Amazon FSx for NetApp ONTAP で SnapMirrorのカスケードをしてみた
本番ワークロードのONTAPからSnapMirrorで移行するのは避けたい
こんにちは、のんピ(@non____97)です。
皆さんはオンプレミスONTAPからAmazon FSx for NetApp(以降FSxN)に以降する際に、本番ワークロードのONTAPを触るのは避けたいなと思ったことはありますか? 私はあります。
本番ワークロードに対して何か設定変更を加えたことによって、負荷がかかってしまい業務に影響があるのは避けたいところです。
おおよそ、重要なワークロードについてはSnapMirror/SnapVaultでSnapshotコピーを別のONTAPに転送していると考えられます。
そのような場面では転送先のONTAPからFSxNに対して、さらにSnapMirrorでSnapshotコピーを転送したいものです。果たしてそんなことはできるのでしょうか。
はい、SnapMirrorのカスケード構成を組むことで対応可能です。
抜粋 : How to configure SnapMirror cascade relationship - NetApp Knowledge Base
SnapMirrorのカスケードを行うことで、ソースボリュームのパフォーマンスを低下させずに、別のストレージにSnapshotコピーを転送することができます。
実際に試してみました。
いきなりまとめ
- カスケードSnapMirrorでは
snapmirror protect
を使用することができない - カスケードSnapMirrorでも重複排除やaggregateレイヤーのデータ削減効果を維持できる
- 2つ目のボリュームで実行したStorage Efficiencyによる重複排除効果を3つ目以降のボリュームに維持するためには、
転送元ボリューム -> 2つ目のボリューム -> 3つ目のボリューム
で2回SnapMirrorによる転送を行う必要がある - 2つ目のボリュームで実行したInactive data compressionなどaggregateレイヤーのデータ削減効果を3つ目以降のボリュームに維持するためには、
2つ目のボリューム -> 3つ目のボリューム
でSnapMirrorを転送するだけでよい- ただし、転送元ボリュームのTiering PolicyがAllの場合は、aggregateレイヤーのデータ削減効果を維持できない
- カスケードSnapMirrorの中間のボリュームが使用できなくなった場合、直接転送元ボリュームから最終的な転送先ボリューム間で同期を行うことができる
- 中間ボリュームと最終的な転送先ボリューム間のSnapMirror relastionshipを削除した上で、転送元ボリュームと最終的な転送先ボリューム間でSnapMirror relasionshipの作成をし、
snapmirror resync
による再同期が必要 - 再同期する際、共通のSnapshot以降に保持しているデータは削除される
- 中間ボリュームと最終的な転送先ボリューム間のSnapMirror relastionshipを削除した上で、転送元ボリュームと最終的な転送先ボリューム間でSnapMirror relasionshipの作成をし、
検証環境
検証環境の構成は以下のとおりです。
今回は同一FSxNファイルシステム内で完結していますが、それぞれ別のFSxNファイルシステム間でも問題なく動作します。
その際のクラスターピアリングやSVMピアリングはソースクラスターと最終的な転送先のクラスター間で行う必要はありません。直接SnapMirrorで転送するクラスター間でのみクラスターピアリングやSVMピアリングを行えば良いです。
SnapMirrorでは、SnapMirrorデスティネーションから別のシステムにデータをレプリケートできます。したがって、あるSnapMirror関係のデスティネーションであるシステムを、別のSnapMirror関係のソースとして使用できます。これは、1つのサイトから複数のサイトにデータを分散する場合に便利です。これをカスケードと呼びます。カスケードトポロジでは、プライマリクラスタとセカンダリクラスタの間、およびセカンダリクラスタとターシャリクラスタの間に、クラスタ間ネットワークを作成する必要があります。プライマリクラスタとターシャリクラスタの間にクラスタ間ネットワークは必要ありません。に、2個のホップから成るカスケード構成の例を示します。
SVMは以下の3つを用意しました。
- svm
- svm2
- svm3
::> vserver show Admin Operational Root Vserver Type Subtype State State Volume Aggregate ----------- ------- ---------- ---------- ----------- ---------- ---------- svm data default running running svm_root aggr1 svm2 data default running running svm2_root aggr1 svm3 data default running running svm3_root aggr1 3 entries were displayed.
ボリュームは以下のとおりです。
::> volume show Vserver Volume Aggregate State Type Size Available Used% --------- ------------ ------------ ---------- ---- ---------- ---------- ----- svm svm_root aggr1 online RW 1GB 972.3MB 0% svm vol1 aggr1 online RW 16GB 15.20GB 0% svm2 svm2_root aggr1 online RW 1GB 972.5MB 0% svm3 svm3_root aggr1 online RW 1GB 972.5MB 0% 4 entries were displayed.
テスト用ファイルを書き込むvol1
はTiering Policy Noneで、Storage Efficiencyは無効化しています。これはファイルを作成するタイミングで重複排除や圧縮などがかかるのを防ぐためです。
::> volume show -volume vol1 -fields tiering-policy vserver volume tiering-policy ------- ------ -------------- svm vol1 none ::> volume efficiency show -volume vol1 Vserver Volume State Status Progress Policy ---------- ---------------- --------- ----------- ------------------ ---------- svm vol1 Disabled Idle Idle for 00:02:38 auto ::> set diag Warning: These diagnostic commands are for use by NetApp personnel only. Do you want to continue? {y|n}: y ::*> volume efficiency show -volume vol1* -fields state, policy, storage-efficiency-mode, inline-compression, inline-dedupe, compression, data-compaction, auto-adaptive-compression-savings, auto-adaptive-compression-existing-volume, using-auto-adaptive-compression vserver volume state policy compression inline-compression storage-efficiency-mode inline-dedupe data-compaction auto-adaptive-compression-savings using-auto-adaptive-compression auto-adaptive-compression-existing-volume ------- ------ -------- ------ ----------- ------------------ ----------------------- ------------- --------------- --------------------------------- ------------------------------- ----------------------------------------- svm vol1 Disabled auto false false efficient false false true false false
やってみる
テスト用ファイルの作成
テスト用のファイルを作成します
作成するファイルは以下の3つです。
1
で埋めた1GiBのテキストファイル- ランダムブロックの1GiBのバイナリファイル
- 2のコピー
1つ目のファイルは重複排除や圧縮が効くのかどうかを確認するためで、2つ目と3つ目のファイルは重複排除が効くのかを確認するためのファイルです。
作成時のログは以下のとおりです。
$ sudo mount -t nfs svm-0058ae83d258ab2e3.fs-0ab6f9b00824a187c.fsx.us-east-1.amazonaws.com:/vol1 /mnt/fsxn/vol1 $ df -hT -t nfs4 Filesystem Type Size Used Avail Use% Mounted on svm-0058ae83d258ab2e3.fs-0ab6f9b00824a187c.fsx.us-east-1.amazonaws.com:/vol1 nfs4 16G 320K 16G 1% /mnt/fsxn/vol1 $ head -c 1G /dev/zero | tr \\0 1 | sudo tee /mnt/fsxn/vol1/1_padding_file > /dev/null $ ls -lh /mnt/fsxn/vol1/ total 1.1G -rw-r--r--. 1 root root 1.0G Dec 22 01:46 1_padding_file $ sudo dd if=/dev/urandom of=/mnt/fsxn/vol1/urandom_block_file bs=1M count=1024 1024+0 records in 1024+0 records out 1073741824 bytes (1.1 GB, 1.0 GiB) copied, 6.25093 s, 172 MB/s $ sudo cp /mnt/fsxn/vol1/urandom_block_file /mnt/fsxn/vol1/urandom_block_file_copy $ ls -lh /mnt/fsxn/vol1/ total 3.1G -rw-r--r--. 1 root root 1.0G Dec 22 01:46 1_padding_file -rw-r--r--. 1 root root 1.0G Dec 22 01:47 urandom_block_file -rw-r--r--. 1 root root 1.0G Dec 22 01:47 urandom_block_file_copy $ df -hT -t nfs4 Filesystem Type Size Used Avail Use% Mounted on svm-0058ae83d258ab2e3.fs-0ab6f9b00824a187c.fsx.us-east-1.amazonaws.com:/vol1 nfs4 16G 3.1G 13G 20% /mnt/fsxn/vol1
Storage Efficiencyを無効化しているので、重複排除は行われず3.1GB消費されていることが分かります。
ファイル作成後、ONTAP CLIからボリュームやaggregateの情報を確認します。
::*> volume show -volume vol1* -fields available, filesystem-size, total, used, percent-used, performance-tier-inactive-user-data, performance-tier-inactive-user-data-percent, size, dedupe-space-saved, dedupe-space-saved-percent, dedupe-space-shared,logical-used, logical-used-percent,logical-used-by-afs, logical-available vserver volume size available filesystem-size total used percent-used dedupe-space-saved dedupe-space-saved-percent dedupe-space-shared logical-used logical-used-percent logical-available logical-used-by-afs performance-tier-inactive-user-data performance-tier-inactive-user-data-percent ------- ------ ---- --------- --------------- ------- ------ ------------ ------------------ -------------------------- ------------------- ------------ -------------------- ----------------- ------------------- ----------------------------------- ------------------------------------------- svm vol1 16GB 12.19GB 16GB 15.20GB 3.01GB 19% 0B 0% 0B 3.01GB 20% - 3.01GB 0B 0% ::*> volume show-footprint -volume vol1 Vserver : svm Volume : vol1 Feature Used Used% -------------------------------- ---------- ----- Volume Data Footprint 3.01GB 0% Footprint in Performance Tier 3.02GB 100% Footprint in FSxFabricpoolObjectStore 0B 0% Volume Guarantee 0B 0% Flexible Volume Metadata 92.66MB 0% Delayed Frees 4.99MB 0% File Operation Metadata 4KB 0% Total Footprint 3.11GB 0% Effective Total Footprint 3.11GB 0% ::*> aggr show-efficiency -instance Name of the Aggregate: aggr1 Node where Aggregate Resides: FsxId0ab6f9b00824a187c-01 Logical Size Used by Volumes, Clones, Snapshot Copies in the Aggregate: 3.01GB Total Physical Used: 2.13GB Total Storage Efficiency Ratio: 1.42:1 Total Data Reduction Logical Used Without Snapshots: 3.01GB Total Data Reduction Physical Used Without Snapshots: 2.13GB Total Data Reduction Efficiency Ratio Without Snapshots: 1.42:1 Total Data Reduction Logical Used without snapshots and flexclones: 3.01GB Total Data Reduction Physical Used without snapshots and flexclones: 2.13GB Total Data Reduction Efficiency Ratio without snapshots and flexclones: 1.42:1 Total Logical Size Used by Volumes, Clones, Snapshot Copies in the FabricPool Performance Tier: 3.02GB Total Physical Used in FabricPool Performance Tier: 2.14GB Total FabricPool Performance Tier Storage Efficiency Ratio: 1.41:1 Total Data Reduction Logical Used without snapshots and flexclones in the FabricPool Performance Tier: 3.01GB Total Data Reduction Physical Used without snapshots and flexclones in the FabricPool Performance Tier: 2.13GB Total FabricPool Performance Tier Data Reduction Efficiency Ratio without snapshots and flexclones: 1.41:1 Logical Space Used for All Volumes: 3.01GB Physical Space Used for All Volumes: 3.01GB Space Saved by Volume Deduplication: 0B Space Saved by Volume Deduplication and pattern detection: 0B Volume Deduplication Savings ratio: 1.00:1 Space Saved by Volume Compression: 0B Volume Compression Savings ratio: 1.00:1 Space Saved by Inline Zero Pattern Detection: 0B Volume Data Reduction SE Ratio: 1.00:1 Logical Space Used by the Aggregate: 3.12GB Physical Space Used by the Aggregate: 2.13GB Space Saved by Aggregate Data Reduction: 1017MB Aggregate Data Reduction SE Ratio: 1.47:1 Logical Size Used by Snapshot Copies: 1.66MB Physical Size Used by Snapshot Copies: 708KB Snapshot Volume Data Reduction Ratio: 2.40:1 Logical Size Used by FlexClone Volumes: 0B Physical Sized Used by FlexClone Volumes: 0B FlexClone Volume Data Reduction Ratio: 1.00:1 Snapshot And FlexClone Volume Data Reduction SE Ratio: 2.40:1 Number of Volumes Offline: 0 Number of SIS Disabled Volumes: 3 Number of SIS Change Log Disabled Volumes: 1
重複排除は行われていないようです。ただし、Space Saved by Aggregate Data Reduction: 1017MB
であることからaggregateレイヤーで何かしらのデータ削減が行われていそうです。Storage Efficiencyを無効にしているため、圧縮やコンパクションも効いていつもりですが何でしょうか。
vol1 の Storage Efficiency有効化
テストファイルを作成したので、vol1
でStorage Efficiencyを有効化します。
::*> volume efficiency on -vserver svm -volume vol1 Efficiency for volume "vol1" of Vserver "svm" is enabled. ::*> volume efficiency show -volume vol1* -fields state, policy, storage-efficiency-mode, inline-compression, inline-dedupe, compression, data-compaction, auto-adaptive-compression-savings, auto-adaptive-compression-existing-volume, using-auto-adaptive-compression vserver volume state policy compression inline-compression storage-efficiency-mode inline-dedupe data-compaction auto-adaptive-compression-savings using-auto-adaptive-compression auto-adaptive-compression-existing-volume ------- ------ ------- ------ ----------- ------------------ ----------------------- ------------- --------------- --------------------------------- ------------------------------- ----------------------------------------- svm vol1 Enabled auto false false efficient false false true false false
vol1 の Inactive data compression有効化
vol1
でInactive data compressionを有効化します。
Inactive data compressionを有効化する際はusing-auto-adaptive-compression
がtrueである必要があるので、事前にcompression
をtrueを指定します。
::*> volume efficiency inactive-data-compression show Vserver Volume Is-Enabled Scan Mode Progress Status Compression-Algorithm ---------- ------ ---------- --------- -------- ------ --------------------- svm vol1 false - IDLE SUCCESS lzopro ::*> volume efficiency modify -vserver svm -volume vol1 -compression true ::*> volume efficiency show -volume vol1* -fields state, policy, storage-efficiency-mode, inline-compression, inline-dedupe, compression, data-compaction, auto-adaptive-compression-savings, auto-adaptive-compression-existing-volume, using-auto-adaptive-compression vserver volume state policy compression inline-compression storage-efficiency-mode inline-dedupe data-compaction auto-adaptive-compression-savings using-auto-adaptive-compression auto-adaptive-compression-existing-volume ------- ------ ------- ------ ----------- ------------------ ----------------------- ------------- --------------- --------------------------------- ------------------------------- ----------------------------------------- svm vol1 Enabled auto false true efficient false true true true false ::*> volume efficiency inactive-data-compression showVserver Volume Is-Enabled Scan Mode Progress Status Compression-Algorithm ---------- ------ ---------- --------- -------- ------ --------------------- svm vol1 false - IDLE SUCCESS lzopro ::*> volume efficiency inactive-data-compression modify -vserver svm -volume vol1 -is-enabled true ::*> volume efficiency inactive-data-compression show Vserver Volume Is-Enabled Scan Mode Progress Status Compression-Algorithm ---------- ------ ---------- --------- -------- ------ --------------------- svm vol1 true - IDLE SUCCESS lzopro
SVMピアリングの作成
SVMピアリングを行います。
作成するSVMピアリングは以下の2つです。
- svm - svm2
- svm2 - svm3
::*> vserver peer create -vserver svm -peer-vserver svm2 -applications snapmirror Info: 'vserver peer create' command is successful. ::*> vserver peer create -vserver svm2 -peer-vserver svm3 -applications snapmirror Info: 'vserver peer create' command is successful. ::*> vserver peer show Peer Peer Peering Remote Vserver Vserver State Peer Cluster Applications Vserver ----------- ----------- ------------ ----------------- -------------- --------- svm svm2 peered FsxId0ab6f9b00824a187c snapmirror svm2 svm2 svm peered FsxId0ab6f9b00824a187c snapmirror svm svm2 svm3 peered FsxId0ab6f9b00824a187c snapmirror svm3 svm3 svm2 peered FsxId0ab6f9b00824a187c snapmirror svm2 4 entries were displayed.
svm と svm2 のボリューム間のSnapMirror initialize
svm と svm2 のボリューム間のSnapMirror relationshipのinitializeを行います。
::*> snapmirror protect -path-list svm:vol1 -destination-vserver svm2 -policy MirrorAllSnapshots -auto-initialize true -support-tiering true -tiering-policy none [Job 57] Job is queued: snapmirror protect for list of source endpoints beginning with "svm:vol1". ::*> snapmirror show Progress Source Destination Mirror Relationship Total Last Path Type Path State Status Progress Healthy Updated ----------- ---- ------------ ------- -------------- --------- ------- -------- svm:vol1 XDP svm2:vol1_dst Uninitialized Transferring 1.74GB true 12/22 01:56:30 ::*> snapmirror show Progress Source Destination Mirror Relationship Total Last Path Type Path State Status Progress Healthy Updated ----------- ---- ------------ ------- -------------- --------- ------- -------- svm:vol1 XDP svm2:vol1_dst Uninitialized Finalizing 2.06GB true 12/22 01:56:45 ::*> snapmirror show Progress Source Destination Mirror Relationship Total Last Path Type Path State Status Progress Healthy Updated ----------- ---- ------------ ------- -------------- --------- ------- -------- svm:vol1 XDP svm2:vol1_dst Snapmirrored Idle - true - ::*> snapmirror show -instance Source Path: svm:vol1 Source Cluster: - Source Vserver: svm Source Volume: vol1 Destination Path: svm2:vol1_dst Destination Cluster: - Destination Vserver: svm2 Destination Volume: vol1_dst Relationship Type: XDP Relationship Group Type: none Managing Vserver: svm2 SnapMirror Schedule: - SnapMirror Policy Type: async-mirror SnapMirror Policy: MirrorAllSnapshots Tries Limit: - Throttle (KB/sec): unlimited Consistency Group Item Mappings: - Current Transfer Throttle (KB/sec): - Mirror State: Snapmirrored Relationship Status: Idle File Restore File Count: - File Restore File List: - Transfer Snapshot: - Snapshot Progress: - Total Progress: - Network Compression Ratio: - Snapshot Checkpoint: - Newest Snapshot: snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_015623 Newest Snapshot Timestamp: 12/22 01:56:24 Exported Snapshot: snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_015623 Exported Snapshot Timestamp: 12/22 01:56:24 Healthy: true Relationship ID: 4f726f26-a06d-11ee-981e-bdd56ead09c8 Source Vserver UUID: 04ee1778-a058-11ee-981e-bdd56ead09c8 Destination Vserver UUID: 5af907bb-a065-11ee-981e-bdd56ead09c8 Current Operation ID: - Transfer Type: - Transfer Error: - Last Transfer Type: update Last Transfer Error: - Last Transfer Error Codes: - Last Transfer Size: 0B Last Transfer Network Compression Ratio: 1:1 Last Transfer Duration: 0:0:0 Last Transfer From: svm:vol1 Last Transfer End Timestamp: 12/22 01:56:59 Unhealthy Reason: - Progress Last Updated: - Relationship Capability: 8.2 and above Lag Time: 0:0:57 Current Transfer Priority: - SMTape Operation: - Destination Volume Node Name: FsxId0ab6f9b00824a187c-01 Identity Preserve Vserver DR: - Number of Successful Updates: 1 Number of Failed Updates: 0 Number of Successful Resyncs: 0 Number of Failed Resyncs: 0 Number of Successful Breaks: 0 Number of Failed Breaks: 0 Total Transfer Bytes: 2229164624 Total Transfer Time in Seconds: 36 Source Volume MSIDs Preserved: - OpMask: ffffffffffffffff Is Auto Expand Enabled: - Percent Complete for Current Status: -
initializeが完了しました。
Storage Efficiency、ボリューム、aggregate、Snapshotの情報を確認します。
::*> volume efficiency show -volume vol1* -fields state, policy, storage-efficiency-mode, inline-compression, inline-dedupe, compression, data-compaction, auto-adaptive-compression-savings, auto-adaptive-compression-existing-volume, using-auto-adaptive-compression vserver volume state policy compression inline-compression storage-efficiency-mode inline-dedupe data-compaction auto-adaptive-compression-savings using-auto-adaptive-compression auto-adaptive-compression-existing-volume ------- ------ ------- ------ ----------- ------------------ ----------------------- ------------- --------------- --------------------------------- ------------------------------- ----------------------------------------- svm vol1 Enabled auto false true efficient false true true true false svm2 vol1_dst Disabled - false true efficient false true true true false 2 entries were displayed. ::*> volume show -volume vol1* -fields available, filesystem-size, total, used, percent-used, performance-tier-inactive-user-data, performance-tier-inactive-user-data-percent, size, dedupe-space-saved, dedupe-space-saved-percent, dedupe-space-shared,logical-used, logical-used-percent,logical-used-by-afs, logical-available vserver volume size available filesystem-size total used percent-used dedupe-space-saved dedupe-space-saved-percent dedupe-space-shared logical-used logical-used-percent logical-available logical-used-by-afs performance-tier-inactive-user-data performance-tier-inactive-user-data-percent ------- ------ ---- --------- --------------- ------- ------ ------------ ------------------ -------------------------- ------------------- ------------ -------------------- ----------------- ------------------- ----------------------------------- ------------------------------------------- svm vol1 16GB 12.19GB 16GB 15.20GB 3.01GB 19% 0B 0% 0B 3.01GB 20% - 3.01GB 0B 0% svm2 vol1_dst 3.82GB 615.1MB 3.82GB 3.63GB 3.03GB 83% 0B 0% 3GB 3.03GB 83% - 3.03GB 0B 0% 2 entries were displayed. ::*> volume show-footprint -volume vol1_dst Vserver : svm2 Volume : vol1_dst Feature Used Used% -------------------------------- ---------- ----- Volume Data Footprint 3.03GB 0% Footprint in Performance Tier 3.06GB 100% Footprint in FSxFabricpoolObjectStore 0B 0% Volume Guarantee 0B 0% Flexible Volume Metadata 30.29MB 0% Delayed Frees 30.31MB 0% File Operation Metadata 4KB 0% Total Footprint 3.09GB 0% Effective Total Footprint 3.09GB 0% ::*> aggr show-efficiency -instance Name of the Aggregate: aggr1 Node where Aggregate Resides: FsxId0ab6f9b00824a187c-01 Logical Size Used by Volumes, Clones, Snapshot Copies in the Aggregate: 12.08GB Total Physical Used: 4.18GB Total Storage Efficiency Ratio: 2.89:1 Total Data Reduction Logical Used Without Snapshots: 6.04GB Total Data Reduction Physical Used Without Snapshots: 4.18GB Total Data Reduction Efficiency Ratio Without Snapshots: 1.44:1 Total Data Reduction Logical Used without snapshots and flexclones: 6.04GB Total Data Reduction Physical Used without snapshots and flexclones: 4.18GB Total Data Reduction Efficiency Ratio without snapshots and flexclones: 1.44:1 Total Logical Size Used by Volumes, Clones, Snapshot Copies in the FabricPool Performance Tier: 12.08GB Total Physical Used in FabricPool Performance Tier: 4.20GB Total FabricPool Performance Tier Storage Efficiency Ratio: 2.88:1 Total Data Reduction Logical Used without snapshots and flexclones in the FabricPool Performance Tier: 6.04GB Total Data Reduction Physical Used without snapshots and flexclones in the FabricPool Performance Tier: 4.19GB Total FabricPool Performance Tier Data Reduction Efficiency Ratio without snapshots and flexclones: 1.44:1 Logical Space Used for All Volumes: 6.04GB Physical Space Used for All Volumes: 6.04GB Space Saved by Volume Deduplication: 0B Space Saved by Volume Deduplication and pattern detection: 0B Volume Deduplication Savings ratio: 1.00:1 Space Saved by Volume Compression: 0B Volume Compression Savings ratio: 1.00:1 Space Saved by Inline Zero Pattern Detection: 0B Volume Data Reduction SE Ratio: 1.00:1 Logical Space Used by the Aggregate: 6.17GB Physical Space Used by the Aggregate: 4.18GB Space Saved by Aggregate Data Reduction: 1.99GB Aggregate Data Reduction SE Ratio: 1.47:1 Logical Size Used by Snapshot Copies: 6.04GB Physical Size Used by Snapshot Copies: 1.01MB Snapshot Volume Data Reduction Ratio: 6138.38:1 Logical Size Used by FlexClone Volumes: 0B Physical Sized Used by FlexClone Volumes: 0B FlexClone Volume Data Reduction Ratio: 1.00:1 Snapshot And FlexClone Volume Data Reduction SE Ratio: 6138.38:1 Number of Volumes Offline: 0 Number of SIS Disabled Volumes: 3 Number of SIS Change Log Disabled Volumes: 1 ::*> snapshot show -volume vol1* ---Blocks--- Vserver Volume Snapshot Size Total% Used% -------- -------- ------------------------------------- -------- ------ ----- svm vol1 snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_015623 152KB 0% 0% svm2 vol1_dst snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_015623 148KB 0% 0% 2 entries were displayed.
svm2にvol1_dst
というボリュームが作成されています。転送元ボリュームであるvol1
はStorage Efficiencyが有効ですが、こちらのボリュームはStorage Efficiencyが無効になっています。
また、Space Saved by Aggregate Data Reduction
は1.99GB
と1GB弱増えていました。このことからSnapMirrorがaggregateレイヤーでのデータ削減効果を維持してくれていることが分かります。
vol1_dst で Storage Efficiencyを有効化
svm2 のボリュームであるvol1_dst
でStorage Efficiencyを有効化します。
::*> volume efficiency on -vserver svm2 -volume vol1_dst Efficiency for volume "vol1_dst" of Vserver "svm2" is enabled. ::*> volume efficiency show -volume vol1* -fields state, policy, storage-efficiency-mode, inline-compression, inline-dedupe, compression, data-compaction, auto-adaptive-compression-savings, auto-adaptive-compression-existing-volume, using-auto-adaptive-compression vserver volume state policy compression inline-compression storage-efficiency-mode inline-dedupe data-compaction auto-adaptive-compression-savings using-auto-adaptive-compression auto-adaptive-compression-existing-volume ------- ------ ------- ------ ----------- ------------------ ----------------------- ------------- --------------- --------------------------------- ------------------------------- ----------------------------------------- svm vol1 Enabled auto false true efficient false true true true false svm2 vol1_dst Enabled - false true efficient false true true true false 2 entries were displayed.
vol1_dst で Inactive data compressionを実行
svm2のボリュームであるvol1_dst
でInactive data compressionを実行します。
重複排除よりも先に圧縮したのは、圧縮効果をわかりやすくするためです。仮に重複排除後に圧縮をすると、1
で埋め尽くしているテキストファイルは重複排除によりほとんどのデータが削減されてしまい、圧縮が行われたのか分かりづらくなってしまうのではないかと考えます。
::*> volume efficiency inactive-data-compression start -vserver svm2 -volume vol1_dst -inactive-days 0 Inactive data compression scan started on volume "vol1_dst" in Vserver "svm2" ::*> volume efficiency inactive-data-compression show -volume vol1_dst -instance Volume: vol1_dst Vserver: svm2 Is Enabled: true Scan Mode: - Progress: IDLE Status: SUCCESS Compression Algorithm: lzopro Failure Reason: - Total Blocks: - Total blocks Processed: - Percentage: - Phase1 L1s Processed: - Phase1 Lns Skipped: - Phase2 Total Blocks: - Phase2 Blocks Processed: - Number of Cold Blocks Encountered: 526232 Number of Repacked Blocks: 0 Number of Compression Done Blocks: 160 Number of Vol-Overwrites: 0 Time since Last Inactive Data Compression Scan started(sec): 22 Time since Last Inactive Data Compression Scan ended(sec): 21 Time since Last Successful Inactive Data Compression Scan started(sec): - Time since Last Successful Inactive Data Compression Scan ended(sec): 21 Average time for Cold Data Compression(sec): 0 Tuning Enabled: true Threshold: 14 Threshold Upper Limit: 21 Threshold Lower Limit: 14 Client Read history window: 14 Incompressible Data Percentage: 66%
Number of Compression Done Blocks: 160
であることからあまり圧縮されていませんね。
注目すべきポイントはIncompressible Data Percentage: 66%
になっているところです。これは全体の66%のデータを圧縮できなかったという意味です。3GiB書き込んだうちの2GiBがランダムなデータブロックのバイナリファイルです。そして、以下記事でも確認しているとおり/dev/urandom
で生成したランダムなデータブロックのバイナリファイルはInactive data compressionではほとんど圧縮できません。
そのため、おそらく1
で埋めたテキストファイルは圧縮済みで圧縮しなかったのではないかと推測します。
Storage Efficiency、ボリューム、aggregate、Snapshotの情報を確認します。
::*> volume efficiency show -volume vol1* -fields state, policy, storage-efficiency-mode, inline-compression, inline-dedupe, compression, data-compaction, auto-adaptive-compression-savings, auto-adaptive-compression-existing-volume, using-auto-adaptive-compression vserver volume state policy compression inline-compression storage-efficiency-mode inline-dedupe data-compaction auto-adaptive-compression-savings using-auto-adaptive-compression auto-adaptive-compression-existing-volume ------- ------ ------- ------ ----------- ------------------ ----------------------- ------------- --------------- --------------------------------- ------------------------------- ----------------------------------------- svm vol1 Enabled auto false true efficient false true true true false svm2 vol1_dst Enabled - false true efficient false true true true false 2 entries were displayed. ::*> volume show -volume vol1* -fields available, filesystem-size, total, used, percent-used, performance-tier-inactive-user-data, performance-tier-inactive-user-data-percent, size, dedupe-space-saved, dedupe-space-saved-percent, dedupe-space-shared, logical-used, logical-used-percent,logical-used-by-afs, logical-available vserver volume size available filesystem-size total used percent-used dedupe-space-saved dedupe-space-saved-percent dedupe-space-shared logical-used logical-used-percent logical-available logical-used-by-afs performance-tier-inactive-user-data performance-tier-inactive-user-data-percent ------- ------ ---- --------- --------------- ------- ------ ------------ ------------------ -------------------------- ------------------- ------------ -------------------- ----------------- ------------------- ----------------------------------- ------------------------------------------- svm vol1 16GB 12.19GB 16GB 15.20GB 3.01GB 19% 0B 0% 0B 3.01GB 20% - 3.01GB 0B 0% svm2 vol1_dst 3.82GB 615.0MB 3.82GB 3.63GB 3.03GB 83% 0B 0% 3GB 3.03GB 83% - 3.03GB 0B 0% 2 entries were displayed. ::*> volume show-footprint -volume vol1_dst Vserver : svm2 Volume : vol1_dst Feature Used Used% -------------------------------- ---------- ----- Volume Data Footprint 3.03GB 0% Footprint in Performance Tier 3.06GB 100% Footprint in FSxFabricpoolObjectStore 0B 0% Volume Guarantee 0B 0% Flexible Volume Metadata 30.29MB 0% Delayed Frees 31.34MB 0% File Operation Metadata 4KB 0% Total Footprint 3.09GB 0% Footprint Data Reduction 1.01GB 0% Auto Adaptive Compression 1.01GB 0% Effective Total Footprint 2.08GB 0% ::*> aggr show-efficiency -instance Name of the Aggregate: aggr1 Node where Aggregate Resides: FsxId0ab6f9b00824a187c-01 Logical Size Used by Volumes, Clones, Snapshot Copies in the Aggregate: 12.06GB Total Physical Used: 4.25GB Total Storage Efficiency Ratio: 2.84:1 Total Data Reduction Logical Used Without Snapshots: 6.02GB Total Data Reduction Physical Used Without Snapshots: 4.25GB Total Data Reduction Efficiency Ratio Without Snapshots: 1.42:1 Total Data Reduction Logical Used without snapshots and flexclones: 6.02GB Total Data Reduction Physical Used without snapshots and flexclones: 4.25GB Total Data Reduction Efficiency Ratio without snapshots and flexclones: 1.42:1 Total Logical Size Used by Volumes, Clones, Snapshot Copies in the FabricPool Performance Tier: 12.09GB Total Physical Used in FabricPool Performance Tier: 4.29GB Total FabricPool Performance Tier Storage Efficiency Ratio: 2.82:1 Total Data Reduction Logical Used without snapshots and flexclones in the FabricPool Performance Tier: 6.04GB Total Data Reduction Physical Used without snapshots and flexclones in the FabricPool Performance Tier: 4.29GB Total FabricPool Performance Tier Data Reduction Efficiency Ratio without snapshots and flexclones: 1.41:1 Logical Space Used for All Volumes: 6.02GB Physical Space Used for All Volumes: 6.02GB Space Saved by Volume Deduplication: 0B Space Saved by Volume Deduplication and pattern detection: 0B Volume Deduplication Savings ratio: 1.00:1 Space Saved by Volume Compression: 0B Volume Compression Savings ratio: 1.00:1 Space Saved by Inline Zero Pattern Detection: 0B Volume Data Reduction SE Ratio: 1.00:1 Logical Space Used by the Aggregate: 6.23GB Physical Space Used by the Aggregate: 4.25GB Space Saved by Aggregate Data Reduction: 1.99GB Aggregate Data Reduction SE Ratio: 1.47:1 Logical Size Used by Snapshot Copies: 6.04GB Physical Size Used by Snapshot Copies: 2.25MB Snapshot Volume Data Reduction Ratio: 2751.27:1 Logical Size Used by FlexClone Volumes: 0B Physical Sized Used by FlexClone Volumes: 0B FlexClone Volume Data Reduction Ratio: 1.00:1 Snapshot And FlexClone Volume Data Reduction SE Ratio: 2751.27:1 Number of Volumes Offline: 0 Number of SIS Disabled Volumes: 3 Number of SIS Change Log Disabled Volumes: 0 ::*> snapshot show -volume vol1* ---Blocks--- Vserver Volume Snapshot Size Total% Used% -------- -------- ------------------------------------- -------- ------ ----- svm vol1 snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_015623 152KB 0% 0% svm2 vol1_dst snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_015623 160KB 0% 0% 2 entries were displayed.
Space Saved by Aggregate Data Reduction
の結果は変わっていませんが、volume show-footprint
のAuto Adaptive Compression
が1.01GBとなっています。
このことから元々aggregateレイヤーで圧縮が行われていたと推測されます。
vol1_dst で Storage Efficiencyを実行
次にsvm2のボリュームであるvol1_dst
でStorage Efficiencyを実行します。
::*> volume efficiency show -volume vol1* -fields changelog-usage, changelog-size, logical-data-size, state, progress, last-op-size vserver volume state progress last-op-size changelog-usage changelog-size logical-data-size ------- ------ ------- ----------------- ------------ --------------- -------------- ----------------- svm vol1 Enabled Idle for 02:55:16 0B 0% 0B 3.01GB svm2 vol1_dst Enabled Idle for 00:08:51 0B 0% 0B 3.03GB 2 entries were displayed. ::*> volume efficiency start -vserver svm2 -volume vol1_dst -scan-old-data Warning: This operation scans all of the data in volume "vol1_dst" of Vserver "svm2". It might take a significant time, and degrade performance during that time. Do you want to continue? {y|n}: y The efficiency operation for volume "vol1_dst" of Vserver "svm2" has started. ::*> volume efficiency show -volume vol1* -fields changelog-usage, changelog-size, logical-data-size, state, progress, last-op-size vserver volume state progress last-op-size changelog-usage changelog-size logical-data-size ------- ------ ------- ----------------- ------------ --------------- -------------- ----------------- svm vol1 Enabled Idle for 02:55:36 0B 0% 0B 3.01GB svm2 vol1_dst Enabled 186368 KB Scanned 0B 0% 0B 3.03GB 2 entries were displayed. ::*> volume efficiency show -volume vol1* -fields changelog-usage, changelog-size, logical-data-size, state, progress, last-op-size vserver volume state progress last-op-size changelog-usage changelog-size logical-data-size ------- ------ ------- ----------------- ------------ --------------- -------------- ----------------- svm vol1 Enabled Idle for 02:55:49 0B 0% 0B 3.01GB svm2 vol1_dst Enabled 2635776 KB Scanned 0B 0% 0B 3.03GB 2 entries were displayed. ::*> volume efficiency show -volume vol1* -fields changelog-usage, changelog-size, logical-data-size, state, progress, last-op-size vserver volume state progress last-op-size changelog-usage changelog-size logical-data-size ------- ------ ------- ----------------- ------------ --------------- -------------- ----------------- svm vol1 Enabled Idle for 02:56:04 0B 0% 0B 3.01GB svm2 vol1_dst Enabled 5515264 KB Scanned 0B 0% 0B 3.03GB 2 entries were displayed. ::*> volume efficiency show -volume vol1* -fields changelog-usage, changelog-size, logical-data-size, state, progress, last-op-size vserver volume state progress last-op-size changelog-usage changelog-size logical-data-size ------- ------ ------- ----------------- ------------ --------------- -------------- ----------------- svm vol1 Enabled Idle for 02:57:11 0B 0% 0B 3.01GB svm2 vol1_dst Enabled 3387724 KB (64%) Done 0B 0% 120KB 3.03GB 2 entries were displayed. ::*> volume efficiency show -volume vol1* -fields changelog-usage, changelog-size, logical-data-size, state, progress, last-op-size vserver volume state progress last-op-size changelog-usage changelog-size logical-data-size ------- ------ ------- ----------------- ------------ --------------- -------------- ----------------- svm vol1 Enabled Idle for 02:57:58 0B 0% 0B 3.01GB svm2 vol1_dst Enabled Idle for 00:00:18 6GB 0% 420KB 3.03GB 2 entries were displayed.
6GB分操作をしたようです。logical-data-size
にあるとおり、ボリュームに書き込まれたデータサイズは3GBほどであるのに、6GB分処理したのは謎です。
ボリューム、aggregate、Snapshotの情報を確認します。
::*> volume show -volume vol1* -fields available, filesystem-size, total, used, percent-used, performance-tier-inactive-user-data, performance-tier-inactive-user-data-percent, size, dedupe-space-saved, dedupe-space-saved-percent, dedupe-space-shared, logical-used, logical-used-percent,logical-used-by-afs, logical-available vserver volume size available filesystem-size total used percent-used dedupe-space-saved dedupe-space-saved-percent dedupe-space-shared logical-used logical-used-percent logical-available logical-used-by-afs performance-tier-inactive-user-data performance-tier-inactive-user-data-percent ------- ------ ---- --------- --------------- ------- ------ ------------ ------------------ -------------------------- ------------------- ------------ -------------------- ----------------- ------------------- ----------------------------------- ------------------------------------------- svm vol1 16GB 12.19GB 16GB 15.20GB 3.01GB 19% 0B 0% 0B 3.01GB 20% - 3.01GB 0B 0% svm2 vol1_dst 3.58GB 541.2MB 3.58GB 3.40GB 2.87GB 84% 1.89GB 40% 1.06GB 4.75GB 140% - 3.02GB 0B 0% 2 entries were displayed. ::*> volume show-footprint -volume vol1_dst Vserver : svm2 Volume : vol1_dst Feature Used Used% -------------------------------- ---------- ----- Volume Data Footprint 3.05GB 0% Footprint in Performance Tier 3.08GB 100% Footprint in FSxFabricpoolObjectStore 0B 0% Volume Guarantee 0B 0% Flexible Volume Metadata 30.29MB 0% Deduplication Metadata 6.02MB 0% Deduplication 6.02MB 0% Delayed Frees 26.04MB 0% File Operation Metadata 4KB 0% Total Footprint 3.11GB 0% Footprint Data Reduction 1.02GB 0% Auto Adaptive Compression 1.02GB 0% Effective Total Footprint 2.09GB 0% ::*> aggr show-efficiency -instance Name of the Aggregate: aggr1 Node where Aggregate Resides: FsxId0ab6f9b00824a187c-01 Logical Size Used by Volumes, Clones, Snapshot Copies in the Aggregate: 12.06GB Total Physical Used: 4.41GB Total Storage Efficiency Ratio: 2.73:1 Total Data Reduction Logical Used Without Snapshots: 6.02GB Total Data Reduction Physical Used Without Snapshots: 3.09GB Total Data Reduction Efficiency Ratio Without Snapshots: 1.95:1 Total Data Reduction Logical Used without snapshots and flexclones: 6.02GB Total Data Reduction Physical Used without snapshots and flexclones: 3.09GB Total Data Reduction Efficiency Ratio without snapshots and flexclones: 1.95:1 Total Logical Size Used by Volumes, Clones, Snapshot Copies in the FabricPool Performance Tier: 12.09GB Total Physical Used in FabricPool Performance Tier: 4.46GB Total FabricPool Performance Tier Storage Efficiency Ratio: 2.71:1 Total Data Reduction Logical Used without snapshots and flexclones in the FabricPool Performance Tier: 6.05GB Total Data Reduction Physical Used without snapshots and flexclones in the FabricPool Performance Tier: 3.14GB Total FabricPool Performance Tier Data Reduction Efficiency Ratio without snapshots and flexclones: 1.92:1 Logical Space Used for All Volumes: 6.02GB Physical Space Used for All Volumes: 4.12GB Space Saved by Volume Deduplication: 1.89GB Space Saved by Volume Deduplication and pattern detection: 1.89GB Volume Deduplication Savings ratio: 1.46:1 Space Saved by Volume Compression: 0B Volume Compression Savings ratio: 1.00:1 Space Saved by Inline Zero Pattern Detection: 0B Volume Data Reduction SE Ratio: 1.46:1 Logical Space Used by the Aggregate: 6.40GB Physical Space Used by the Aggregate: 4.41GB Space Saved by Aggregate Data Reduction: 1.99GB Aggregate Data Reduction SE Ratio: 1.45:1 Logical Size Used by Snapshot Copies: 6.04GB Physical Size Used by Snapshot Copies: 1.91GB Snapshot Volume Data Reduction Ratio: 3.16:1 Logical Size Used by FlexClone Volumes: 0B Physical Sized Used by FlexClone Volumes: 0B FlexClone Volume Data Reduction Ratio: 1.00:1 Snapshot And FlexClone Volume Data Reduction SE Ratio: 3.16:1 Number of Volumes Offline: 0 Number of SIS Disabled Volumes: 3 Number of SIS Change Log Disabled Volumes: 0 ::*> snapshot show -volume vol1* ---Blocks--- Vserver Volume Snapshot Size Total% Used% -------- -------- ------------------------------------- -------- ------ ----- svm vol1 snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_015623 152KB 0% 0% svm2 vol1_dst snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_015623 1.91GB 53% 63% 2 entries were displayed.
重複排除量が1.89GBとなっていますね。
ただし、Total Physical Used
が4.25GBから4.41GBに僅かならが増えています。そのため、重複排除したデータブロックは物理的には解放されていないことが分かります。
svm3 に SnapMirrorの転送先ボリュームの作成
svm3にSnapMirrorの転送先ボリュームを作成します。
カスケードSnapMirrorではsnapmirror protectをサポートしていません。
snapmirror protect SnapMirrorカスケードの作成 で は サポートされていません
そのため、以下の操作は手動で行う必要があります。
- カスケードSnapMirrorの転送先ボリュームの作成
- SnapMirror relationshipの作成
- SnapMirror initializeの実行
ボリュームはvol1_dst_dst
という名前で作成します。
::*> volume create -vserver svm3 -volume vol1_dst_dst -aggregate aggr1 -state online -type DP -size 4GB -tiering-policy none [Job 63] Job succeeded: Successful ::*> volume show -volume vol1* -fields type, autosize-mode, max-autosize vserver volume max-autosize autosize-mode type ------- ------ ------------ ------------- ---- svm vol1 19.20GB off RW svm2 vol1_dst 100TB grow_shrink DP svm3 vol1_dst_dst 100TB grow_shrink DP 3 entries were displayed.
作成したボリュームのStorage Efficiencyを有効化します。
::*> volume efficiency show -volume vol1* -fields state, policy, storage-efficiency-mode, inline-compression, inline-dedupe, compression, data-compaction, auto-adaptive-compression-savings, auto-adaptive-compression-existing-volume, using-auto-adaptive-compression vserver volume state policy compression inline-compression storage-efficiency-mode inline-dedupe data-compaction auto-adaptive-compression-savings using-auto-adaptive-compression auto-adaptive-compression-existing-volume ------- ------ ------- ------ ----------- ------------------ ----------------------- ------------- --------------- --------------------------------- ------------------------------- ----------------------------------------- svm vol1 Enabled auto false true efficient false true true true false svm2 vol1_dst Enabled - false true efficient false true true true false svm3 vol1_dst_dst Disabled - false false - false true false false false 3 entries were displayed. ::*> volume efficiency inactive-data-compression show -volume vol1_dst_dst -instance There are no entries matching your query. ::*> volume efficiency on -vserver svm3 -volume vol1_dst_dst Efficiency for volume "vol1_dst_dst" of Vserver "svm3" is enabled. ::*> volume efficiency show -volume vol1* -fields state, policy, storage-efficiency-mode, inline-compression, inline-dedupe, compression, data-compaction, auto-adaptive-compression-savings, auto-adaptive-compression-existing-volume, using-auto-adaptive-compression vserver volume state policy compression inline-compression storage-efficiency-mode inline-dedupe data-compaction auto-adaptive-compression-savings using-auto-adaptive-compression auto-adaptive-compression-existing-volume ------- ------ ------- ------ ----------- ------------------ ----------------------- ------------- --------------- --------------------------------- ------------------------------- ----------------------------------------- svm vol1 Enabled auto false true efficient false true true true false svm2 vol1_dst Enabled - false true efficient false true true true false svm3 vol1_dst_dst Enabled - false false - false true false false false 3 entries were displayed. ::*> volume efficiency inactive-data-compression show -volume vol1_dst_dst -instance There are no entries matching your query.
svm2 と svm3 のボリューム間のSnapMirror initialize
svm2 と svm3 のボリューム間のSnapMirror relationshipのinitializeを行います。
まず、SnapMirror relasionshipの作成をします。
::*> snapmirror create -source-path svm2:vol1_dst -destination-vserver svm3 -destination-volume vol1_dst_dst -policy MirrorAllSnapshots Operation succeeded: snapmirror create for the relationship with destination "svm3:vol1_dst_dst". ::*> snapmirror show Progress Source Destination Mirror Relationship Total Last Path Type Path State Status Progress Healthy Updated ----------- ---- ------------ ------- -------------- --------- ------- -------- svm:vol1 XDP svm2:vol1_dst Snapmirrored Idle - true - svm2:vol1_dst XDP svm3:vol1_dst_dst Uninitialized Idle - true - 2 entries were displayed.
次にinitializeを行います。
::*> snapmirror initialize -destination-path svm3:vol1_dst_dst -source-path svm2:vol1_dst Operation is queued: snapmirror initialize of destination "svm3:vol1_dst_dst". ::*> snapmirror show Progress Source Destination Mirror Relationship Total Last Path Type Path State Status Progress Healthy Updated ----------- ---- ------------ ------- -------------- --------- ------- -------- svm:vol1 XDP svm2:vol1_dst Snapmirrored Idle - true - svm2:vol1_dst XDP svm3:vol1_dst_dst Uninitialized Transferring 0B true 12/22 04:44:26 2 entries were displayed. ::*> snapmirror show Progress Source Destination Mirror Relationship Total Last Path Type Path State Status Progress Healthy Updated ----------- ---- ------------ ------- -------------- --------- ------- -------- svm:vol1 XDP svm2:vol1_dst Snapmirrored Idle - true - svm2:vol1_dst XDP svm3:vol1_dst_dst Uninitialized Transferring 376.5MB true 12/22 04:44:39 2 entries were displayed. ::*> snapmirror show Progress Source Destination Mirror Relationship Total Last Path Type Path State Status Progress Healthy Updated ----------- ---- ------------ ------- -------------- --------- ------- -------- svm:vol1 XDP svm2:vol1_dst Snapmirrored Idle - true - svm2:vol1_dst XDP svm3:vol1_dst_dst Snapmirrored Idle - true - 2 entries were displayed. ::*> snapmirror show -destination-path svm3:vol1_dst_dst Source Path: svm2:vol1_dst Source Cluster: - Source Vserver: svm2 Source Volume: vol1_dst Destination Path: svm3:vol1_dst_dst Destination Cluster: - Destination Vserver: svm3 Destination Volume: vol1_dst_dst Relationship Type: XDP Relationship Group Type: none Managing Vserver: svm3 SnapMirror Schedule: - SnapMirror Policy Type: async-mirror SnapMirror Policy: MirrorAllSnapshots Tries Limit: - Throttle (KB/sec): unlimited Consistency Group Item Mappings: - Current Transfer Throttle (KB/sec): - Mirror State: Snapmirrored Relationship Status: Idle File Restore File Count: - File Restore File List: - Transfer Snapshot: - Snapshot Progress: - Total Progress: - Network Compression Ratio: - Snapshot Checkpoint: - Newest Snapshot: snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_015623 Newest Snapshot Timestamp: 12/22 01:56:24 Exported Snapshot: snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_015623 Exported Snapshot Timestamp: 12/22 01:56:24 Healthy: true Relationship ID: b0b2694d-a084-11ee-981e-bdd56ead09c8 Source Vserver UUID: 5af907bb-a065-11ee-981e-bdd56ead09c8 Destination Vserver UUID: 674509fa-a065-11ee-981e-bdd56ead09c8 Current Operation ID: - Transfer Type: - Transfer Error: - Last Transfer Type: update Last Transfer Error: - Last Transfer Error Codes: - Last Transfer Size: 0B Last Transfer Network Compression Ratio: 1:1 Last Transfer Duration: 0:0:0 Last Transfer From: svm2:vol1_dst Last Transfer End Timestamp: 12/22 04:45:00 Unhealthy Reason: - Progress Last Updated: - Relationship Capability: 8.2 and above Lag Time: 2:49:31 Current Transfer Priority: - SMTape Operation: - Destination Volume Node Name: FsxId0ab6f9b00824a187c-01 Identity Preserve Vserver DR: - Number of Successful Updates: 1 Number of Failed Updates: 0 Number of Successful Resyncs: 0 Number of Failed Resyncs: 0 Number of Successful Breaks: 0 Number of Failed Breaks: 0 Total Transfer Bytes: 2229260985 Total Transfer Time in Seconds: 34 Source Volume MSIDs Preserved: - OpMask: ffffffffffffffff Is Auto Expand Enabled: - Percent Complete for Current Status: -
Storage Efficiency、ボリューム、aggregate、Snapshotの情報を確認します。
::*> volume efficiency show -volume vol1* -fields state, policy, storage-efficiency-mode, inline-compression, inline-dedupe, compression, data-compaction, auto-adaptive-compression-savings, auto-adaptive-compression-existing-volume, using-auto-adaptive-compression vserver volume state policy compression inline-compression storage-efficiency-mode inline-dedupe data-compaction auto-adaptive-compression-savings using-auto-adaptive-compression auto-adaptive-compression-existing-volume ------- ------ ------- ------ ----------- ------------------ ----------------------- ------------- --------------- --------------------------------- ------------------------------- ----------------------------------------- svm vol1 Enabled auto false true efficient false true true true false svm2 vol1_dst Enabled - false true efficient false true true true false svm3 vol1_dst_dst Enabled - false true efficient false true true true false 3 entries were displayed. ::*> volume efficiency inactive-data-compression show -volume vol1_dst_dst -instance Volume: vol1_dst_dst Vserver: svm3 Is Enabled: true Scan Mode: - Progress: IDLE Status: SUCCESS Compression Algorithm: lzopro Failure Reason: - Total Blocks: - Total blocks Processed: - Percentage: - Phase1 L1s Processed: - Phase1 Lns Skipped: - Phase2 Total Blocks: - Phase2 Blocks Processed: - Number of Cold Blocks Encountered: 0 Number of Repacked Blocks: 0 Number of Compression Done Blocks: 0 Number of Vol-Overwrites: 0 Time since Last Inactive Data Compression Scan started(sec): 0 Time since Last Inactive Data Compression Scan ended(sec): 0 Time since Last Successful Inactive Data Compression Scan started(sec): - Time since Last Successful Inactive Data Compression Scan ended(sec): 0 Average time for Cold Data Compression(sec): 0 Tuning Enabled: true Threshold: 14 Threshold Upper Limit: 21 Threshold Lower Limit: 14 Client Read history window: 14 Incompressible Data Percentage: 0% ::*> volume show -volume vol1* -fields available, filesystem-size, total, used, percent-used, performance-tier-inactive-user-data, performance-tier-inactive-user-data-percent, size, dedupe-space-saved, dedupe-space-saved-percent, dedupe-space-shared, logical-used, logical-used-percent,logical-used-by-afs, logical-available vserver volume size available filesystem-size total used percent-used dedupe-space-saved dedupe-space-saved-percent dedupe-space-shared logical-used logical-used-percent logical-available logical-used-by-afs performance-tier-inactive-user-data performance-tier-inactive-user-data-percent ------- ------ ---- --------- --------------- ------- ------ ------------ ------------------ -------------------------- ------------------- ------------ -------------------- ----------------- ------------------- ----------------------------------- ------------------------------------------- svm vol1 16GB 12.19GB 16GB 15.20GB 3.01GB 19% 0B 0% 0B 3.01GB 20% - 3.01GB 0B 0% svm2 vol1_dst 3.58GB 541.2MB 3.58GB 3.40GB 2.87GB 84% 1.89GB 40% 1.06GB 4.75GB 140% - 3.02GB 0B 0% svm3 vol1_dst_dst 4GB 954.2MB 4GB 4GB 3.07GB 76% 0B 0% 3GB 3.07GB 77% - 3.07GB 0B 0% 3 entries were displayed. ::*> volume show-footprint -volume vol1_dst_dst Vserver : svm3 Volume : vol1_dst_dst Feature Used Used% -------------------------------- ---------- ----- Volume Data Footprint 3.07GB 0% Footprint in Performance Tier 3.10GB 100% Footprint in FSxFabricpoolObjectStore 0B 0% Volume Guarantee 0B 0% Flexible Volume Metadata 31.23MB 0% Delayed Frees 34.50MB 0% File Operation Metadata 4KB 0% Total Footprint 3.13GB 0% Effective Total Footprint 3.13GB 0% ::*> aggr show-efficiency -instance Name of the Aggregate: aggr1 Node where Aggregate Resides: FsxId0ab6f9b00824a187c-01 Logical Size Used by Volumes, Clones, Snapshot Copies in the Aggregate: 18.16GB Total Physical Used: 6.33GB Total Storage Efficiency Ratio: 2.87:1 Total Data Reduction Logical Used Without Snapshots: 9.04GB Total Data Reduction Physical Used Without Snapshots: 5.03GB Total Data Reduction Efficiency Ratio Without Snapshots: 1.80:1 Total Data Reduction Logical Used without snapshots and flexclones: 9.04GB Total Data Reduction Physical Used without snapshots and flexclones: 5.03GB Total Data Reduction Efficiency Ratio without snapshots and flexclones: 1.80:1 Total Logical Size Used by Volumes, Clones, Snapshot Copies in the FabricPool Performance Tier: 18.23GB Total Physical Used in FabricPool Performance Tier: 6.43GB Total FabricPool Performance Tier Storage Efficiency Ratio: 2.83:1 Total Data Reduction Logical Used without snapshots and flexclones in the FabricPool Performance Tier: 9.12GB Total Data Reduction Physical Used without snapshots and flexclones in the FabricPool Performance Tier: 5.13GB Total FabricPool Performance Tier Data Reduction Efficiency Ratio without snapshots and flexclones: 1.78:1 Logical Space Used for All Volumes: 9.04GB Physical Space Used for All Volumes: 7.15GB Space Saved by Volume Deduplication: 1.89GB Space Saved by Volume Deduplication and pattern detection: 1.89GB Volume Deduplication Savings ratio: 1.26:1 Space Saved by Volume Compression: 0B Volume Compression Savings ratio: 1.00:1 Space Saved by Inline Zero Pattern Detection: 0B Volume Data Reduction SE Ratio: 1.26:1 Logical Space Used by the Aggregate: 9.31GB Physical Space Used by the Aggregate: 6.33GB Space Saved by Aggregate Data Reduction: 2.98GB Aggregate Data Reduction SE Ratio: 1.47:1 Logical Size Used by Snapshot Copies: 9.11GB Physical Size Used by Snapshot Copies: 1.91GB Snapshot Volume Data Reduction Ratio: 4.76:1 Logical Size Used by FlexClone Volumes: 0B Physical Sized Used by FlexClone Volumes: 0B FlexClone Volume Data Reduction Ratio: 1.00:1 Snapshot And FlexClone Volume Data Reduction SE Ratio: 4.76:1 Number of Volumes Offline: 0 Number of SIS Disabled Volumes: 3 Number of SIS Change Log Disabled Volumes: 0 ::*> snapshot show -volume vol1* ---Blocks--- Vserver Volume Snapshot Size Total% Used% -------- -------- ------------------------------------- -------- ------ ----- svm vol1 snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_015623 152KB 0% 0% svm2 vol1_dst snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_015623 1.91GB 53% 63% svm3 vol1_dst_dst snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_015623 144KB 0% 0% 3 entries were displayed.
重複排除が効いているとvol1_dst
から転送しているにも関わらず、vol1_dst_dst
では重複排除が一切効いていないようです。
vol1_dst_dst
上のSnapshotを確認すると、vol1
、vol1_dst_dst
と全く同じものでした。vol1_dst
とvol1_dst_dst
のSnapMirror relationshipのSnapMirror policyはMirrorAllSnapshots
です。こちらのSnapMirror policyにはsm_created
のルールが存在しています。
::*> snapmirror policy show Vserver Policy Policy Number Transfer Name Name Type Of Rules Tries Priority Comment ------- ------------------ ------ -------- ----- -------- ---------- FsxId06a2871837f70a7fc Asynchronous mirror-vault 3 8 normal A unified Asynchronous SnapMirror and SnapVault policy for mirroring the latest active file system and daily and weekly Snapshot copies with an hourly transfer schedule. SnapMirror Label: sm_created Keep: 1 daily 7 weekly 52 Total Keep: 60 FsxId06a2871837f70a7fc AutomatedFailOver automated-failover 1 8 normal Policy for SnapMirror Synchronous with zero RTO guarantee where client I/O will not be disrupted on replication failure. SnapMirror Label: sm_created Keep: 1 Total Keep: 1 FsxId06a2871837f70a7fc CloudBackupDefault vault 1 8 normal Vault policy with daily rule. SnapMirror Label: daily Keep: 7 Total Keep: 7 FsxId06a2871837f70a7fc Continuous continuous 0 8 normal Policy for S3 bucket mirroring. SnapMirror Label: - Keep: - Total Keep: 0 FsxId06a2871837f70a7fc DPDefault async-mirror 2 8 normal Asynchronous SnapMirror policy for mirroring all Snapshot copies and the latest active file system. SnapMirror Label: sm_created Keep: 1 all_source_snapshots 1 Total Keep: 2 FsxId06a2871837f70a7fc DailyBackup vault 1 8 normal Vault policy with a daily rule and a daily transfer schedule. SnapMirror Label: daily Keep: 7 Total Keep: 7 FsxId06a2871837f70a7fc Migrate migrate 2 8 normal Policy for Migrate SnapMirror Label: sm_created Keep: 1 all_source_snapshots 1 Total Keep: 2 FsxId06a2871837f70a7fc MirrorAllSnapshots async-mirror 2 8 normal Asynchronous SnapMirror policy for mirroring all Snapshot copies and the latest active file system. SnapMirror Label: sm_created Keep: 1 all_source_snapshots 1 Total Keep: 2 FsxId06a2871837f70a7fc MirrorAllSnapshotsDiscardNetwork async-mirror 2 8 normal Asynchronous SnapMirror policy for mirroring all Snapshot copies and the latest active file system excluding the network configurations. Discard Configs: network SnapMirror Label: sm_created Keep: 1 all_source_snapshots 1 Total Keep: 2 FsxId06a2871837f70a7fc MirrorAndVault mirror-vault 3 8 normal A unified Asynchronous SnapMirror and SnapVault policy for mirroring the latest active file system and daily and weekly Snapshot copies. SnapMirror Label: sm_created Keep: 1 daily 7 weekly 52 Total Keep: 60 FsxId06a2871837f70a7fc MirrorAndVaultDiscardNetwork mirror-vault 3 8 normal A unified Asynchronous SnapMirror and SnapVault policy for mirroring the latest active file system and daily and weekly Snapshot copies excluding the network configurations. Discard Configs: network SnapMirror Label: sm_created Keep: 1 daily 7 weekly 52 Total Keep: 60 Vserver Policy Policy Number Transfer Name Name Type Of Rules Tries Priority Comment ------- ------------------ ------ -------- ----- -------- ---------- FsxId06a2871837f70a7fc MirrorLatest async-mirror 1 8 normal Asynchronous SnapMirror policy for mirroring the latest active file system. SnapMirror Label: sm_created Keep: 1 Total Keep: 1 FsxId06a2871837f70a7fc SnapCenterSync sync-mirror 2 8 normal Policy for SnapMirror Synchronous for Snap Center with Application Created Snapshot configuration. SnapMirror Label: sm_created Keep: 2 app_consistent 1 Total Keep: 3 FsxId06a2871837f70a7fc StrictSync strict-sync-mirror 1 8 normal Policy for SnapMirror Synchronous where client access will be disrupted on replication failure. SnapMirror Label: sm_created Keep: 2 Total Keep: 2 FsxId06a2871837f70a7fc Sync sync-mirror 1 8 normal Policy for SnapMirror Synchronous where client access will not be disrupted on replication failure. SnapMirror Label: sm_created Keep: 2 Total Keep: 2 FsxId06a2871837f70a7fc Unified7year mirror-vault 4 8 normal Unified SnapMirror policy with 7year retention. SnapMirror Label: sm_created Keep: 1 daily 7 weekly 52 monthly 84 Total Keep: 144 FsxId06a2871837f70a7fc XDPDefault vault 2 8 normal Vault policy with daily and weekly rules. SnapMirror Label: daily Keep: 7 weekly 52 Total Keep: 59 17 entries were displayed.
そのため、「SnapMirror時にSnapshotを作成してくれるのかな?」と思ったのですが、そういう訳ではないようです。
これはSnapMirrorの転送元ボリュームのボリュームタイプがRW
ではなく、DP
であるためです。DP
のボリュームでSnapshotを作成しようとすると、以下のようにエラーになります。
::*> snapshot create -vserver svm2 -volume vol1_dst -snapshot test.2023-12-22_0459 -snapmirror-label test Error: command failed: Snapshot copies can only be created on read/write (RW) volumes
Space Saved by Aggregate Data Reduction: 2.98GB
とSnapMirror前後で1GB増加しています。そのため、SnapMirrorのカスケードにおいてもaggregateレイヤーのデータ削減効果は維持されると考えて良さそうです。
また、無効だったInactive data compressionは有効化されていますね。
カスケードSnapMirrorの再転送
テスト用ファイルの追加
カスケードSnapMirrorの再転送をしてみましょう。
下準備としてテスト用ファイルを追加します。
作成するファイルは以下の3つです。
a
で埋めた1GiBのテキストファイル- ランダムブロックの1GiBのバイナリファイル
- 2のコピー
1つ目のファイルは重複排除や圧縮が効くのかどうかを確認するためで、2つ目と3つ目のファイルは重複排除が効くのかを確認するためのファイルです。
作成時のログは以下のとおりです。
$ sudo dd if=/dev/urandom of=/mnt/fsxn/vol1/urandom_block_file2 bs=1M count=1024 1024+0 records in 1024+0 records out 1073741824 bytes (1.1 GB, 1.0 GiB) copied, 6.38572 s, 168 MB/s $ sudo cp /mnt/fsxn/vol1/urandom_block_file2 /mnt/fsxn/vol1/urandom_block_file2_copy $ head -c 1G /dev/zero | tr \\0 a | sudo tee /mnt/fsxn/vol1/a_padding_file > /dev/null $ ls -lh /mnt/fsxn/vol1 total 6.1G -rw-r--r--. 1 root root 1.0G Dec 22 01:46 1_padding_file -rw-r--r--. 1 root root 1.0G Dec 22 05:28 a_padding_file -rw-r--r--. 1 root root 1.0G Dec 22 01:47 urandom_block_file -rw-r--r--. 1 root root 1.0G Dec 22 05:02 urandom_block_file2 -rw-r--r--. 1 root root 1.0G Dec 22 05:02 urandom_block_file2_copy -rw-r--r--. 1 root root 1.0G Dec 22 01:47 urandom_block_file_copy $ df -hT -t nfs4 Filesystem Type Size Used Avail Use% Mounted on svm-0058ae83d258ab2e3.fs-0ab6f9b00824a187c.fsx.us-east-1.amazonaws.com:/vol1 nfs4 16G 5.4G 9.9G 35% /mnt/fsxn/vol1
Storage Efficiency、ボリューム、aggregateの情報を確認します。
::*> volume show -volume vol1* -fields available, filesystem-size, total, used, percent-used, performance-tier-inactive-user-data, performance-tier-inactive-user-data-percent, size, dedupe-space-saved, dedupe-space-saved-percent, dedupe-space-shared, logical-used, logical-used-percent,logical-used-by-afs, logical-available vserver volume size available filesystem-size total used percent-used dedupe-space-saved dedupe-space-saved-percent dedupe-space-shared logical-used logical-used-percent logical-available logical-used-by-afs performance-tier-inactive-user-data performance-tier-inactive-user-data-percent ------- ------ ---- --------- --------------- ------- ------ ------------ ------------------ -------------------------- ------------------- ------------ -------------------- ----------------- ------------------- ----------------------------------- ------------------------------------------- svm vol1 16GB 9.89GB 16GB 15.20GB 5.31GB 34% 774.5MB 12% 774.5MB 6.06GB 40% - 6.06GB 0B 0% svm2 vol1_dst 3.58GB 541.1MB 3.58GB 3.40GB 2.87GB 84% 1.89GB 40% 1.06GB 4.75GB 140% - 3.02GB 0B 0% svm3 vol1_dst_dst 4GB 954.2MB 4GB 4GB 3.07GB 76% 0B 0% 3GB 3.07GB 77% - 3.07GB 0B 0% 3 entries were displayed. ::*> volume show-footprint -volume vol1 Vserver : svm Volume : vol1 Feature Used Used% -------------------------------- ---------- ----- Volume Data Footprint 5.31GB 1% Footprint in Performance Tier 5.45GB 100% Footprint in FSxFabricpoolObjectStore 0B 0% Volume Guarantee 0B 0% Flexible Volume Metadata 92.66MB 0% Deduplication Metadata 6.02MB 0% Deduplication 6.02MB 0% Delayed Frees 142.8MB 0% File Operation Metadata 4KB 0% Total Footprint 5.54GB 1% Effective Total Footprint 5.54GB 1% ::*> aggr show-efficiency -instance Name of the Aggregate: aggr1 Node where Aggregate Resides: FsxId0ab6f9b00824a187c-01 Logical Size Used by Volumes, Clones, Snapshot Copies in the Aggregate: 21.19GB Total Physical Used: 7.85GB Total Storage Efficiency Ratio: 2.70:1 Total Data Reduction Logical Used Without Snapshots: 12.07GB Total Data Reduction Physical Used Without Snapshots: 6.57GB Total Data Reduction Efficiency Ratio Without Snapshots: 1.84:1 Total Data Reduction Logical Used without snapshots and flexclones: 12.07GB Total Data Reduction Physical Used without snapshots and flexclones: 6.57GB Total Data Reduction Efficiency Ratio without snapshots and flexclones: 1.84:1 Total Logical Size Used by Volumes, Clones, Snapshot Copies in the FabricPool Performance Tier: 21.28GB Total Physical Used in FabricPool Performance Tier: 7.97GB Total FabricPool Performance Tier Storage Efficiency Ratio: 2.67:1 Total Data Reduction Logical Used without snapshots and flexclones in the FabricPool Performance Tier: 12.17GB Total Data Reduction Physical Used without snapshots and flexclones in the FabricPool Performance Tier: 6.70GB Total FabricPool Performance Tier Data Reduction Efficiency Ratio without snapshots and flexclones: 1.81:1 Logical Space Used for All Volumes: 12.07GB Physical Space Used for All Volumes: 9.42GB Space Saved by Volume Deduplication: 2.65GB Space Saved by Volume Deduplication and pattern detection: 2.65GB Volume Deduplication Savings ratio: 1.28:1 Space Saved by Volume Compression: 0B Volume Compression Savings ratio: 1.00:1 Space Saved by Inline Zero Pattern Detection: 0B Volume Data Reduction SE Ratio: 1.28:1 Logical Space Used by the Aggregate: 11.81GB Physical Space Used by the Aggregate: 7.85GB Space Saved by Aggregate Data Reduction: 3.96GB Aggregate Data Reduction SE Ratio: 1.50:1 Logical Size Used by Snapshot Copies: 9.11GB Physical Size Used by Snapshot Copies: 1.91GB Snapshot Volume Data Reduction Ratio: 4.76:1 Logical Size Used by FlexClone Volumes: 0B Physical Sized Used by FlexClone Volumes: 0B FlexClone Volume Data Reduction Ratio: 1.00:1 Snapshot And FlexClone Volume Data Reduction SE Ratio: 4.76:1 Number of Volumes Offline: 0 Number of SIS Disabled Volumes: 3 Number of SIS Change Log Disabled Volumes: 0
774.5MBほど重複排除されています。インライン重複排除は無効にしているので、ポストプロセス重複排除が働いたのでしょう。
また、Space Saved by Aggregate Data Reduction: 3.96GB
とaggregateレイヤーのデータ削減量が1GB増加していました。
svm と svm2 のボリューム間のSnapMirror update
svm と svm2 のボリューム間のSnapMirrorの差分転送を行います。
転送を行うにあたってSnapshotを事前に作成しておきます。意図は特にありません。
::*> snapshot create -vserver svm -volume vol1 -snapshot test.2023-12-22_0533 -snapmirror-label test ::*> snapshot show -volume vol1* ---Blocks--- Vserver Volume Snapshot Size Total% Used% -------- -------- ------------------------------------- -------- ------ ----- svm vol1 snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_015623 228KB 0% 0% test.2023-12-22_0533 156KB 0% 0% svm2 vol1_dst snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_015623 1.91GB 53% 63% svm3 vol1_dst_dst snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_015623 144KB 0% 0% 4 entries were displayed.
それではSnapMirrorの差分転送を行います。
::*> snapmirror update -destination-path svm2:vol1_dst Operation is queued: snapmirror update of destination "svm2:vol1_dst". ::*> snapmirror show Progress Source Destination Mirror Relationship Total Last Path Type Path State Status Progress Healthy Updated ----------- ---- ------------ ------- -------------- --------- ------- -------- svm:vol1 XDP svm2:vol1_dst Snapmirrored Transferring 0B true 12/22 05:35:28 svm2:vol1_dst XDP svm3:vol1_dst_dst Snapmirrored Idle - true - 2 entries were displayed. ::*> volume efficiency show -volume vol1* -fields changelog-usage, changelog-size, logical-data-size, state, progress, last-op-size vserver volume state progress last-op-size changelog-usage changelog-size logical-data-size ------- ------ ------- ----------------- ------------ --------------- -------------- ----------------- svm vol1 Enabled Idle for 00:32:23 1.76GB 7% 12.40MB 6.06GB svm2 vol1_dst Enabled Pre-procesing transfer changelogs 2352608 KB 6GB 0% 448KB 6.11GB svm3 vol1_dst_dst Enabled Idle for 00:54:43 0B 0% 0B 3.07GB 3 entries were displayed. ::*> ::*> volume efficiency show -volume vol1* -fields changelog-usage, changelog-size, logical-data-size, state, progress, last-op-size vserver volume state progress last-op-size changelog-usage changelog-size logical-data-size ------- ------ ------- ----------------- ------------ --------------- -------------- ----------------- svm vol1 Enabled Idle for 00:32:36 1.76GB 7% 12.40MB 6.06GB svm2 vol1_dst Enabled 1245968 KB (95%) Done 6GB 1% 4.30MB 6.15GB svm3 vol1_dst_dst Enabled Idle for 00:54:56 0B 0% 0B 3.07GB 3 entries were displayed. ::*> snapmirror show Progress Source Destination Mirror Relationship Total Last Path Type Path State Status Progress Healthy Updated ----------- ---- ------------ ------- -------------- --------- ------- -------- svm:vol1 XDP svm2:vol1_dst Snapmirrored Idle - true - svm2:vol1_dst XDP svm3:vol1_dst_dst Snapmirrored Idle - true - 2 entries were displayed. ::*> snapmirror show -destination-path svm2:vol1_dst Source Path: svm:vol1 Source Cluster: - Source Vserver: svm Source Volume: vol1 Destination Path: svm2:vol1_dst Destination Cluster: - Destination Vserver: svm2 Destination Volume: vol1_dst Relationship Type: XDP Relationship Group Type: none Managing Vserver: svm2 SnapMirror Schedule: - SnapMirror Policy Type: async-mirror SnapMirror Policy: MirrorAllSnapshots Tries Limit: - Throttle (KB/sec): unlimited Consistency Group Item Mappings: - Current Transfer Throttle (KB/sec): - Mirror State: Snapmirrored Relationship Status: Idle File Restore File Count: - File Restore File List: - Transfer Snapshot: - Snapshot Progress: - Total Progress: - Network Compression Ratio: - Snapshot Checkpoint: - Newest Snapshot: snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_053528 Newest Snapshot Timestamp: 12/22 05:35:28 Exported Snapshot: snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_053528 Exported Snapshot Timestamp: 12/22 05:35:28 Healthy: true Relationship ID: 4f726f26-a06d-11ee-981e-bdd56ead09c8 Source Vserver UUID: 04ee1778-a058-11ee-981e-bdd56ead09c8 Destination Vserver UUID: 5af907bb-a065-11ee-981e-bdd56ead09c8 Current Operation ID: - Transfer Type: - Transfer Error: - Last Transfer Type: update Last Transfer Error: - Last Transfer Error Codes: - Last Transfer Size: 1.33GB Last Transfer Network Compression Ratio: 1:1 Last Transfer Duration: 0:0:12 Last Transfer From: svm:vol1 Last Transfer End Timestamp: 12/22 05:35:40 Unhealthy Reason: - Progress Last Updated: - Relationship Capability: 8.2 and above Lag Time: 0:18:12 Current Transfer Priority: - SMTape Operation: - Destination Volume Node Name: FsxId0ab6f9b00824a187c-01 Identity Preserve Vserver DR: - Number of Successful Updates: 2 Number of Failed Updates: 0 Number of Successful Resyncs: 0 Number of Failed Resyncs: 0 Number of Successful Breaks: 0 Number of Failed Breaks: 0 Total Transfer Bytes: 3658511072 Total Transfer Time in Seconds: 48 Source Volume MSIDs Preserved: - OpMask: ffffffffffffffff Is Auto Expand Enabled: - Percent Complete for Current Status: -
1.33GB転送されたようです。これはテスト用ファイルとして3.0GB追加しましたが、重複排除により700MB、aggregateレイヤーのデータ削減により1GB削減されているためだと考えます。
Storage Efficiency、ボリューム、aggregate、Snapshotの情報を確認します。
::*> volume efficiency show -volume vol1* -fields changelog-usage, changelog-size, logical-data-size, state, progress, last-op-size vserver volume state progress last-op-size changelog-usage changelog-size logical-data-size ------- ------ ------- ----------------- ------------ --------------- -------------- ----------------- svm vol1 Enabled Idle for 00:32:42 1.76GB 7% 12.40MB 6.06GB svm2 vol1_dst Enabled Idle for 00:00:02 45.10MB 1% 4.10MB 6.09GB svm3 vol1_dst_dst Enabled Idle for 00:55:02 0B 0% 0B 3.07GB 3 entries were displayed. ::*> volume efficiency inactive-data-compression show -volume vol1_dst -instance Volume: vol1_dst Vserver: svm2 Is Enabled: true Scan Mode: - Progress: IDLE Status: SUCCESS Compression Algorithm: lzopro Failure Reason: - Total Blocks: - Total blocks Processed: - Percentage: - Phase1 L1s Processed: - Phase1 Lns Skipped: - Phase2 Total Blocks: - Phase2 Blocks Processed: - Number of Cold Blocks Encountered: 0 Number of Repacked Blocks: 0 Number of Compression Done Blocks: 0 Number of Vol-Overwrites: 0 Time since Last Inactive Data Compression Scan started(sec): 41 Time since Last Inactive Data Compression Scan ended(sec): 40 Time since Last Successful Inactive Data Compression Scan started(sec): - Time since Last Successful Inactive Data Compression Scan ended(sec): 40 Average time for Cold Data Compression(sec): 0 Tuning Enabled: true Threshold: 14 Threshold Upper Limit: 21 Threshold Lower Limit: 14 Client Read history window: 14 Incompressible Data Percentage: 66% ::*> volume show -volume vol1* -fields available, filesystem-size, total, used, percent-used, performance-tier-inactive-user-data, performance-tier-inactive-user-data-percent, size, dedupe-space-saved, dedupe-space-saved-percent, dedupe-space-shared, logical-used, logical-used-percent,logical-used-by-afs, logical-available vserver volume size available filesystem-size total used percent-used dedupe-space-saved dedupe-space-saved-percent dedupe-space-shared logical-used logical-used-percent logical-available logical-used-by-afs performance-tier-inactive-user-data performance-tier-inactive-user-data-percent ------- ------ ---- --------- --------------- ------- ------ ------------ ------------------ -------------------------- ------------------- ------------ -------------------- ----------------- ------------------- ----------------------------------- ------------------------------------------- svm vol1 16GB 9.89GB 16GB 15.20GB 5.31GB 34% 774.5MB 12% 774.5MB 6.06GB 40% - 6.06GB 0B 0% svm2 vol1_dst 6.37GB 998.4MB 6.37GB 6.05GB 5.08GB 83% 3.47GB 41% 2.48GB 8.53GB 141% - 6.07GB 0B 0% svm3 vol1_dst_dst 4GB 954.2MB 4GB 4GB 3.07GB 76% 0B 0% 3GB 3.07GB 77% - 3.07GB 0B 0% 3 entries were displayed. ::*> volume show-footprint -volume vol1_dst Vserver : svm2 Volume : vol1_dst Feature Used Used% -------------------------------- ---------- ----- Volume Data Footprint 5.40GB 1% Footprint in Performance Tier 5.45GB 100% Footprint in FSxFabricpoolObjectStore 0B 0% Volume Guarantee 0B 0% Flexible Volume Metadata 43.37MB 0% Deduplication Metadata 12.30MB 0% Deduplication 12.30MB 0% Delayed Frees 53.75MB 0% File Operation Metadata 4KB 0% Total Footprint 5.50GB 1% Footprint Data Reduction 1.80GB 0% Auto Adaptive Compression 1.80GB 0% Effective Total Footprint 3.70GB 0% ::*> aggr show-efficiency -instance Name of the Aggregate: aggr1 Node where Aggregate Resides: FsxId0ab6f9b00824a187c-01 Logical Size Used by Volumes, Clones, Snapshot Copies in the Aggregate: 48.47GB Total Physical Used: 9.25GB Total Storage Efficiency Ratio: 5.24:1 Total Data Reduction Logical Used Without Snapshots: 15.10GB Total Data Reduction Physical Used Without Snapshots: 7.43GB Total Data Reduction Efficiency Ratio Without Snapshots: 2.03:1 Total Data Reduction Logical Used without snapshots and flexclones: 15.10GB Total Data Reduction Physical Used without snapshots and flexclones: 7.43GB Total Data Reduction Efficiency Ratio without snapshots and flexclones: 2.03:1 Total Logical Size Used by Volumes, Clones, Snapshot Copies in the FabricPool Performance Tier: 48.59GB Total Physical Used in FabricPool Performance Tier: 9.42GB Total FabricPool Performance Tier Storage Efficiency Ratio: 5.16:1 Total Data Reduction Logical Used without snapshots and flexclones in the FabricPool Performance Tier: 15.22GB Total Data Reduction Physical Used without snapshots and flexclones in the FabricPool Performance Tier: 7.60GB Total FabricPool Performance Tier Data Reduction Efficiency Ratio without snapshots and flexclones: 2.00:1 Logical Space Used for All Volumes: 15.10GB Physical Space Used for All Volumes: 10.87GB Space Saved by Volume Deduplication: 4.23GB Space Saved by Volume Deduplication and pattern detection: 4.23GB Volume Deduplication Savings ratio: 1.39:1 Space Saved by Volume Compression: 0B Volume Compression Savings ratio: 1.00:1 Space Saved by Inline Zero Pattern Detection: 0B Volume Data Reduction SE Ratio: 1.39:1 Logical Space Used by the Aggregate: 14.17GB Physical Space Used by the Aggregate: 9.25GB Space Saved by Aggregate Data Reduction: 4.92GB Aggregate Data Reduction SE Ratio: 1.53:1 Logical Size Used by Snapshot Copies: 33.38GB Physical Size Used by Snapshot Copies: 2.79GB Snapshot Volume Data Reduction Ratio: 11.98:1 Logical Size Used by FlexClone Volumes: 0B Physical Sized Used by FlexClone Volumes: 0B FlexClone Volume Data Reduction Ratio: 1.00:1 Snapshot And FlexClone Volume Data Reduction SE Ratio: 11.98:1 Number of Volumes Offline: 0 Number of SIS Disabled Volumes: 3 Number of SIS Change Log Disabled Volumes: 0 ::*> snapshot show -volume vol1* ---Blocks--- Vserver Volume Snapshot Size Total% Used% -------- -------- ------------------------------------- -------- ------ ----- svm vol1 snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_015623 228KB 0% 0% test.2023-12-22_0533 160KB 0% 0% snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_053528 148KB 0% 0% svm2 vol1_dst snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_015623 1.91GB 30% 42% test.2023-12-22_0533 893.2MB 14% 25% snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_053528 156KB 0% 0% svm3 vol1_dst_dst snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_015623 144KB 0% 0% 7 entries were displayed.
vol1_dst
の重複排除量が1.89GBから3.47GBと約1.6GB分増えています。
vol1
での重複排除量が約800MBであるので、それを差し引いても800MBほど追加の重複排除が発生しています。これはurandom_block_file2
とurandom_block_file2_copy
間、もしくはa_padding_file
内で重複排除が行われたものと考えられます。
追加の重複排除が行われたのはvol1_dst
のtest.2023-12-22_0533
のサイズが893.2MBであることからも推測できます。
また、Space Saved by Aggregate Data Reduction
が3.96GB から4.92GBと1GBほど増加していました。
vol1_dst で Storage Efficiencyを実行
svm2のボリュームであるvol1_dst
でStorage Efficiencyを実行します。
::*> volume efficiency start -vserver svm2 -volume vol1_dst -scan-old-data Warning: This operation scans all of the data in volume "vol1_dst" of Vserver "svm2". It might take a significant time, and degrade performance during that time. Do you want to continue? {y|n}: y The efficiency operation for volume "vol1_dst" of Vserver "svm2" has started. ::*> volume efficiency show -volume vol1* -fields changelog-usage, changelog-size, logical-data-size, state, progress, last-op-size vserver volume state progress last-op-size changelog-usage changelog-size logical-data-size ------- ------ ------- ----------------- ------------ --------------- -------------- ----------------- svm vol1 Enabled Idle for 00:43:00 1.76GB 7% 12.40MB 6.06GB svm2 vol1_dst Enabled 3194880 KB Scanned 45.10MB 0% 0B 6.05GB svm3 vol1_dst_dst Enabled Idle for 01:05:20 0B 0% 0B 3.07GB 3 entries were displayed. ::*> volume efficiency show -volume vol1* -fields changelog-usage, changelog-size, logical-data-size, state, progress, last-op-size vserver volume state progress last-op-size changelog-usage changelog-size logical-data-size ------- ------ ------- ----------------- ------------ --------------- -------------- ----------------- svm vol1 Enabled Idle for 00:43:14 1.76GB 7% 12.40MB 6.06GB svm2 vol1_dst Enabled 5239752 KB Scanned 45.10MB 0% 0B 6.05GB svm3 vol1_dst_dst Enabled Idle for 01:05:34 0B 0% 0B 3.07GB 3 entries were displayed. ::*> volume efficiency show -volume vol1* -fields changelog-usage, changelog-size, logical-data-size, state, progress, last-op-size vserver volume state progress last-op-size changelog-usage changelog-size logical-data-size ------- ------ ------- ----------------- ------------ --------------- -------------- ----------------- svm vol1 Enabled Idle for 00:43:42 1.76GB 7% 12.40MB 6.06GB svm2 vol1_dst Enabled 4336 KB (0%) Done 45.10MB 0% 0B 6.05GB svm3 vol1_dst_dst Enabled Idle for 01:06:02 0B 0% 0B 3.07GB 3 entries were displayed. ::*> volume efficiency show -volume vol1* -fields changelog-usage, changelog-size, logical-data-size, state, progress, last-op-size vserver volume state progress last-op-size changelog-usage changelog-size logical-data-size ------- ------ ------- ----------------- ------------ --------------- -------------- ----------------- svm vol1 Enabled Idle for 00:45:12 1.76GB 7% 12.40MB 6.06GB svm2 vol1_dst Enabled 6332144 KB (83%) Done 45.10MB 0% 300KB 6.05GB svm3 vol1_dst_dst Enabled Idle for 01:07:32 0B 0% 0B 3.07GB 3 entries were displayed. ::*> volume efficiency show -volume vol1* -fields changelog-usage, changelog-size, logical-data-size, state, progress, last-op-size vserver volume state progress last-op-size changelog-usage changelog-size logical-data-size ------- ------ ------- ----------------- ------------ --------------- -------------- ----------------- svm vol1 Enabled Idle for 00:46:02 1.76GB 7% 12.40MB 6.06GB svm2 vol1_dst Enabled Idle for 00:00:25 9.22GB 0% 420KB 6.06GB svm3 vol1_dst_dst Enabled Idle for 01:08:22 0B 0% 0B 3.07GB 3 entries were displayed.
9.22GB分処理されました。
ボリューム、aggregate、Snapshotの情報を確認します。
::*> volume show -volume vol1* -fields available, filesystem-size, total, used, percent-used, performance-tier-inactive-user-data, performance-tier-inactive-user-data-percent, size, dedupe-space-saved, dedupe-space-saved-percent, dedupe-space-shared, logical-used, logical-used-percent,logical-used-by-afs, logical-available vserver volume size available filesystem-size total used percent-used dedupe-space-saved dedupe-space-saved-percent dedupe-space-shared logical-used logical-used-percent logical-available logical-used-by-afs performance-tier-inactive-user-data performance-tier-inactive-user-data-percent ------- ------ ---- --------- --------------- ------- ------ ------------ ------------------ -------------------------- ------------------- ------------ -------------------- ----------------- ------------------- ----------------------------------- ------------------------------------------- svm vol1 16GB 9.89GB 16GB 15.20GB 5.31GB 34% 774.5MB 12% 774.5MB 6.06GB 40% - 6.06GB 0B 0% svm2 vol1_dst 6.37GB 967.7MB 6.37GB 6.05GB 5.11GB 84% 3.89GB 43% 2.09GB 8.98GB 148% - 6.04GB 0B 0% svm3 vol1_dst_dst 4GB 954.2MB 4GB 4GB 3.07GB 76% 0B 0% 3GB 3.07GB 77% - 3.07GB 0B 0% 3 entries were displayed. ::*> volume show-footprint -volume vol1_dst Vserver : svm2 Volume : vol1_dst Feature Used Used% -------------------------------- ---------- ----- Volume Data Footprint 5.43GB 1% Footprint in Performance Tier 5.49GB 100% Footprint in FSxFabricpoolObjectStore 0B 0% Volume Guarantee 0B 0% Flexible Volume Metadata 43.37MB 0% Deduplication Metadata 12.04MB 0% Deduplication 12.04MB 0% Delayed Frees 61.92MB 0% File Operation Metadata 4KB 0% Total Footprint 5.54GB 1% Footprint Data Reduction 1.81GB 0% Auto Adaptive Compression 1.81GB 0% Effective Total Footprint 3.73GB 0% ::*> aggr show-efficiency -instance Name of the Aggregate: aggr1 Node where Aggregate Resides: FsxId0ab6f9b00824a187c-01 Logical Size Used by Volumes, Clones, Snapshot Copies in the Aggregate: 48.47GB Total Physical Used: 9.33GB Total Storage Efficiency Ratio: 5.20:1 Total Data Reduction Logical Used Without Snapshots: 15.10GB Total Data Reduction Physical Used Without Snapshots: 7.19GB Total Data Reduction Efficiency Ratio Without Snapshots: 2.10:1 Total Data Reduction Logical Used without snapshots and flexclones: 15.10GB Total Data Reduction Physical Used without snapshots and flexclones: 7.19GB Total Data Reduction Efficiency Ratio without snapshots and flexclones: 2.10:1 Total Logical Size Used by Volumes, Clones, Snapshot Copies in the FabricPool Performance Tier: 48.57GB Total Physical Used in FabricPool Performance Tier: 9.47GB Total FabricPool Performance Tier Storage Efficiency Ratio: 5.13:1 Total Data Reduction Logical Used without snapshots and flexclones in the FabricPool Performance Tier: 15.19GB Total Data Reduction Physical Used without snapshots and flexclones in the FabricPool Performance Tier: 7.34GB Total FabricPool Performance Tier Data Reduction Efficiency Ratio without snapshots and flexclones: 2.07:1 Logical Space Used for All Volumes: 15.10GB Physical Space Used for All Volumes: 10.45GB Space Saved by Volume Deduplication: 4.65GB Space Saved by Volume Deduplication and pattern detection: 4.65GB Volume Deduplication Savings ratio: 1.44:1 Space Saved by Volume Compression: 0B Volume Compression Savings ratio: 1.00:1 Space Saved by Inline Zero Pattern Detection: 0B Volume Data Reduction SE Ratio: 1.44:1 Logical Space Used by the Aggregate: 14.25GB Physical Space Used by the Aggregate: 9.33GB Space Saved by Aggregate Data Reduction: 4.92GB Aggregate Data Reduction SE Ratio: 1.53:1 Logical Size Used by Snapshot Copies: 33.38GB Physical Size Used by Snapshot Copies: 3.26GB Snapshot Volume Data Reduction Ratio: 10.25:1 Logical Size Used by FlexClone Volumes: 0B Physical Sized Used by FlexClone Volumes: 0B FlexClone Volume Data Reduction Ratio: 1.00:1 Snapshot And FlexClone Volume Data Reduction SE Ratio: 10.25:1 Number of Volumes Offline: 0 Number of SIS Disabled Volumes: 3 Number of SIS Change Log Disabled Volumes: 0 ::*> snapshot show -volume vol1* ---Blocks--- Vserver Volume Snapshot Size Total% Used% -------- -------- ------------------------------------- -------- ------ ----- svm vol1 snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_015623 228KB 0% 0% test.2023-12-22_0533 160KB 0% 0% snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_053528 148KB 0% 0% svm2 vol1_dst snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_015623 1.91GB 30% 47% test.2023-12-22_0533 893.2MB 14% 29% snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_053528 480.5MB 7% 18% svm3 vol1_dst_dst snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_015623 144KB 0% 0% 7 entries were displayed.
重複排除量は変わりありませんでした。一方でvol1_dst
のsnapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_053528
のサイズが480.5MBとなっていました。内部的にデータブロックの場所が変更されたのではないかと推測します。
また、Auto Adaptive Compressionが1.02GBから1.81GBと増加していました。
vol1_dst で Inactive data compressionを実行
svm2のボリュームであるvol1_dst
でInactive data compressionを実行します。
::*> volume efficiency inactive-data-compression start -vserver svm2 -volume vol1_dst -inactive-days 0 Inactive data compression scan started on volume "vol1_dst" in Vserver "svm2" ::*> volume efficiency inactive-data-compression show -volume vol1_dst -instance Volume: vol1_dst Vserver: svm2 Is Enabled: true Scan Mode: default Progress: RUNNING Status: SUCCESS Compression Algorithm: lzopro Failure Reason: - Total Blocks: - Total blocks Processed: - Percentage: 7% Phase1 L1s Processed: 6168 Phase1 Lns Skipped: L1: 0 L2: 0 L3: 0 L4: 0 L5: 0 L6: 0 L7: 0 Phase2 Total Blocks: 2220400 Phase2 Blocks Processed: 157264 Number of Cold Blocks Encountered: 295168 Number of Repacked Blocks: 0 Number of Compression Done Blocks: 1784 Number of Vol-Overwrites: 0 Time since Last Inactive Data Compression Scan started(sec): 4998 Time since Last Inactive Data Compression Scan ended(sec): 4997 Time since Last Successful Inactive Data Compression Scan started(sec): - Time since Last Successful Inactive Data Compression Scan ended(sec): 4997 Average time for Cold Data Compression(sec): 0 Tuning Enabled: true Threshold: 14 Threshold Upper Limit: 21 Threshold Lower Limit: 14 Client Read history window: 14 Incompressible Data Percentage: 66% ::*> volume efficiency inactive-data-compression show -volume vol1_dst -instance Volume: vol1_dst Vserver: svm2 Is Enabled: true Scan Mode: default Progress: RUNNING Status: SUCCESS Compression Algorithm: lzopro Failure Reason: - Total Blocks: - Total blocks Processed: - Percentage: 72% Phase1 L1s Processed: 6168 Phase1 Lns Skipped: L1: 0 L2: 0 L3: 0 L4: 0 L5: 0 L6: 0 L7: 0 Phase2 Total Blocks: 2220400 Phase2 Blocks Processed: 1608704 Number of Cold Blocks Encountered: 344832 Number of Repacked Blocks: 0 Number of Compression Done Blocks: 10008 Number of Vol-Overwrites: 0 Time since Last Inactive Data Compression Scan started(sec): 5012 Time since Last Inactive Data Compression Scan ended(sec): 5011 Time since Last Successful Inactive Data Compression Scan started(sec): - Time since Last Successful Inactive Data Compression Scan ended(sec): 5011 Average time for Cold Data Compression(sec): 0 Tuning Enabled: true Threshold: 14 Threshold Upper Limit: 21 Threshold Lower Limit: 14 Client Read history window: 14 Incompressible Data Percentage: 66% ::*> volume efficiency inactive-data-compression show -volume vol1_dst -instance Volume: vol1_dst Vserver: svm2 Is Enabled: true Scan Mode: - Progress: IDLE Status: SUCCESS Compression Algorithm: lzopro Failure Reason: - Total Blocks: - Total blocks Processed: - Percentage: - Phase1 L1s Processed: - Phase1 Lns Skipped: - Phase2 Total Blocks: - Phase2 Blocks Processed: - Number of Cold Blocks Encountered: 344832 Number of Repacked Blocks: 0 Number of Compression Done Blocks: 10008 Number of Vol-Overwrites: 0 Time since Last Inactive Data Compression Scan started(sec): 32 Time since Last Inactive Data Compression Scan ended(sec): 11 Time since Last Successful Inactive Data Compression Scan started(sec): - Time since Last Successful Inactive Data Compression Scan ended(sec): 11 Average time for Cold Data Compression(sec): 10 Tuning Enabled: true Threshold: 14 Threshold Upper Limit: 21 Threshold Lower Limit: 14 Client Read history window: 14 Incompressible Data Percentage: 61%
Storage Efficiency、ボリューム、aggregate、Snapshotの情報を確認します。
::*> volume efficiency show -volume vol1* -fields state, policy, storage-efficiency-mode, inline-compression, inline-dedupe, compression, data-compaction, auto-adaptive-compression-savings, auto-adaptive-compression-existing-volume, using-auto-adaptive-compression vserver volume state policy compression inline-compression storage-efficiency-mode inline-dedupe data-compaction auto-adaptive-compression-savings using-auto-adaptive-compression auto-adaptive-compression-existing-volume ------- ------ ------- ------ ----------- ------------------ ----------------------- ------------- --------------- --------------------------------- ------------------------------- ----------------------------------------- svm vol1 Enabled auto false true efficient false true true true false svm2 vol1_dst Enabled - false true efficient false true true true false svm3 vol1_dst_dst Enabled - false true efficient false true true true false 3 entries were displayed. ::*> volume show -volume vol1* -fields available, filesystem-size, total, used, percent-used, performance-tier-inactive-user-data, performance-tier-inactive-user-data-percent, size, dedupe-space-saved, dedupe-space-saved-percent, dedupe-space-shared, logical-used, logical-used-percent,logical-used-by-afs, logical-available vserver volume size available filesystem-size total used percent-used dedupe-space-saved dedupe-space-saved-percent dedupe-space-shared logical-used logical-used-percent logical-available logical-used-by-afs performance-tier-inactive-user-data performance-tier-inactive-user-data-percent ------- ------ ---- --------- --------------- ------- ------ ------------ ------------------ -------------------------- ------------------- ------------ -------------------- ----------------- ------------------- ----------------------------------- ------------------------------------------- svm vol1 16GB 9.89GB 16GB 15.20GB 5.31GB 34% 774.5MB 12% 774.5MB 6.06GB 40% - 6.06GB 0B 0% svm2 vol1_dst 6.37GB 967.6MB 6.37GB 6.05GB 5.11GB 84% 3.89GB 43% 2.09GB 8.98GB 148% - 6.04GB 0B 0% svm3 vol1_dst_dst 4GB 954.2MB 4GB 4GB 3.07GB 76% 0B 0% 3GB 3.07GB 77% - 3.07GB 0B 0% 3 entries were displayed. ::*> volume show-footprint -volume vol1_dst Vserver : svm2 Volume : vol1_dst Feature Used Used% -------------------------------- ---------- ----- Volume Data Footprint 5.43GB 1% Footprint in Performance Tier 5.49GB 100% Footprint in FSxFabricpoolObjectStore 0B 0% Volume Guarantee 0B 0% Flexible Volume Metadata 43.37MB 0% Deduplication Metadata 12.04MB 0% Deduplication 12.04MB 0% Delayed Frees 62.57MB 0% File Operation Metadata 4KB 0% Total Footprint 5.54GB 1% Footprint Data Reduction 2.04GB 0% Auto Adaptive Compression 2.04GB 0% Effective Total Footprint 3.50GB 0% ::*> aggr show-efficiency -instance Name of the Aggregate: aggr1 Node where Aggregate Resides: FsxId0ab6f9b00824a187c-01 Logical Size Used by Volumes, Clones, Snapshot Copies in the Aggregate: 48.41GB Total Physical Used: 9.23GB Total Storage Efficiency Ratio: 5.25:1 Total Data Reduction Logical Used Without Snapshots: 15.03GB Total Data Reduction Physical Used Without Snapshots: 7.11GB Total Data Reduction Efficiency Ratio Without Snapshots: 2.11:1 Total Data Reduction Logical Used without snapshots and flexclones: 15.03GB Total Data Reduction Physical Used without snapshots and flexclones: 7.11GB Total Data Reduction Efficiency Ratio without snapshots and flexclones: 2.11:1 Total Logical Size Used by Volumes, Clones, Snapshot Copies in the FabricPool Performance Tier: 48.57GB Total Physical Used in FabricPool Performance Tier: 9.44GB Total FabricPool Performance Tier Storage Efficiency Ratio: 5.15:1 Total Data Reduction Logical Used without snapshots and flexclones in the FabricPool Performance Tier: 15.19GB Total Data Reduction Physical Used without snapshots and flexclones in the FabricPool Performance Tier: 7.32GB Total FabricPool Performance Tier Data Reduction Efficiency Ratio without snapshots and flexclones: 2.08:1 Logical Space Used for All Volumes: 15.03GB Physical Space Used for All Volumes: 10.39GB Space Saved by Volume Deduplication: 4.65GB Space Saved by Volume Deduplication and pattern detection: 4.65GB Volume Deduplication Savings ratio: 1.45:1 Space Saved by Volume Compression: 0B Volume Compression Savings ratio: 1.00:1 Space Saved by Inline Zero Pattern Detection: 0B Volume Data Reduction SE Ratio: 1.45:1 Logical Space Used by the Aggregate: 14.19GB Physical Space Used by the Aggregate: 9.23GB Space Saved by Aggregate Data Reduction: 4.96GB Aggregate Data Reduction SE Ratio: 1.54:1 Logical Size Used by Snapshot Copies: 33.38GB Physical Size Used by Snapshot Copies: 3.26GB Snapshot Volume Data Reduction Ratio: 10.25:1 Logical Size Used by FlexClone Volumes: 0B Physical Sized Used by FlexClone Volumes: 0B FlexClone Volume Data Reduction Ratio: 1.00:1 Snapshot And FlexClone Volume Data Reduction SE Ratio: 10.25:1 Number of Volumes Offline: 0 Number of SIS Disabled Volumes: 3 Number of SIS Change Log Disabled Volumes: 0 ::*> snapshot show -volume vol1* ---Blocks--- Vserver Volume Snapshot Size Total% Used% -------- -------- ------------------------------------- -------- ------ ----- svm vol1 snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_015623 228KB 0% 0% test.2023-12-22_0533 160KB 0% 0% snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_053528 148KB 0% 0% svm2 vol1_dst snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_015623 1.91GB 30% 47% test.2023-12-22_0533 893.2MB 14% 29% snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_053528 480.5MB 7% 18% svm3 vol1_dst_dst snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_015623 144KB 0% 0% 7 entries were displayed.
Space Saved by Aggregate Data Reduction
が4.92GBから4.96GBと若干増加していました。Auto Adaptive Compression
についても1.81GBから2.04GBと増加していました。
svm2 と svm3 のボリューム間のSnapMirror update
svm と svm2 のボリューム間のSnapMirrorの差分転送を行います。
::*> snapmirror update -destination-path svm3:vol1_dst_dst Operation is queued: snapmirror update of destination "svm3:vol1_dst_dst". ::*> snapmirror show Progress Source Destination Mirror Relationship Total Last Path Type Path State Status Progress Healthy Updated ----------- ---- ------------ ------- -------------- --------- ------- -------- svm:vol1 XDP svm2:vol1_dst Snapmirrored Idle - true - svm2:vol1_dst XDP svm3:vol1_dst_dst Snapmirrored Transferring 642.0MB true 12/22 05:55:29 2 entries were displayed. ::*> snapmirror show Progress Source Destination Mirror Relationship Total Last Path Type Path State Status Progress Healthy Updated ----------- ---- ------------ ------- -------------- --------- ------- -------- svm:vol1 XDP svm2:vol1_dst Snapmirrored Idle - true - svm2:vol1_dst XDP svm3:vol1_dst_dst Snapmirrored Finalizing 1.38GB true 12/22 05:55:44 2 entries were displayed. ::*> snapmirror show Progress Source Destination Mirror Relationship Total Last Path Type Path State Status Progress Healthy Updated ----------- ---- ------------ ------- -------------- --------- ------- -------- svm:vol1 XDP svm2:vol1_dst Snapmirrored Idle - true - svm2:vol1_dst XDP svm3:vol1_dst_dst Snapmirrored Idle - true - 2 entries were displayed. ::*> snapmirror show -destination-path svm3:vol1_dst_dst Source Path: svm2:vol1_dst Source Cluster: - Source Vserver: svm2 Source Volume: vol1_dst Destination Path: svm3:vol1_dst_dst Destination Cluster: - Destination Vserver: svm3 Destination Volume: vol1_dst_dst Relationship Type: XDP Relationship Group Type: none Managing Vserver: svm3 SnapMirror Schedule: - SnapMirror Policy Type: async-mirror SnapMirror Policy: MirrorAllSnapshots Tries Limit: - Throttle (KB/sec): unlimited Consistency Group Item Mappings: - Current Transfer Throttle (KB/sec): - Mirror State: Snapmirrored Relationship Status: Idle File Restore File Count: - File Restore File List: - Transfer Snapshot: - Snapshot Progress: - Total Progress: - Network Compression Ratio: - Snapshot Checkpoint: - Newest Snapshot: snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_053528 Newest Snapshot Timestamp: 12/22 05:35:28 Exported Snapshot: snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_053528 Exported Snapshot Timestamp: 12/22 05:35:28 Healthy: true Relationship ID: b0b2694d-a084-11ee-981e-bdd56ead09c8 Source Vserver UUID: 5af907bb-a065-11ee-981e-bdd56ead09c8 Destination Vserver UUID: 674509fa-a065-11ee-981e-bdd56ead09c8 Current Operation ID: - Transfer Type: - Transfer Error: - Last Transfer Type: update Last Transfer Error: - Last Transfer Error Codes: - Last Transfer Size: 1.39GB Last Transfer Network Compression Ratio: 1:1 Last Transfer Duration: 0:0:44 Last Transfer From: svm2:vol1_dst Last Transfer End Timestamp: 12/22 05:56:10 Unhealthy Reason: - Progress Last Updated: - Relationship Capability: 8.2 and above Lag Time: 0:20:56 Current Transfer Priority: - SMTape Operation: - Destination Volume Node Name: FsxId0ab6f9b00824a187c-01 Identity Preserve Vserver DR: - Number of Successful Updates: 2 Number of Failed Updates: 0 Number of Successful Resyncs: 0 Number of Failed Resyncs: 0 Number of Successful Breaks: 0 Number of Failed Breaks: 0 Total Transfer Bytes: 3720662445 Total Transfer Time in Seconds: 78 Source Volume MSIDs Preserved: - OpMask: ffffffffffffffff Is Auto Expand Enabled: - Percent Complete for Current Status: -
1.39GB転送されたようです。vol1
からvol1_dst
にSnapMirrorの差分同期をした時とほぼ同じサイズですね。
Storage Efficiency、ボリューム、aggregate、Snapshotの情報を確認します。
::*> volume efficiency show -volume vol1* -fields changelog-usage, changelog-size, logical-data-size, state, progress, last-op-size vserver volume state progress last-op-size changelog-usage changelog-size logical-data-size ------- ------ ------- ----------------- ------------ --------------- -------------- ----------------- svm vol1 Enabled Idle for 00:53:29 1.76GB 7% 12.40MB 6.06GB svm2 vol1_dst Enabled Idle for 00:07:52 9.22GB 0% 420KB 6.06GB svm3 vol1_dst_dst Enabled 3353944 KB (100%) Done 0B 0% 780KB 6.27GB 3 entries were displayed. ::*> volume efficiency show -volume vol1* -fields changelog-usage, changelog-size, logical-data-size, state, progress, last-op-sizevserver volume state progress last-op-size changelog-usage changelog-size logical-data-size ------- ------ ------- ----------------- ------------ --------------- -------------- ----------------- svm vol1 Enabled Idle for 00:55:34 1.76GB 7% 12.40MB 6.06GB svm2 vol1_dst Enabled Idle for 00:09:57 9.22GB 0% 420KB 6.06GB svm3 vol1_dst_dst Enabled Idle for 00:02:00 5.24GB 0% 836KB 6.06GB 3 entries were displayed. ::*> volume efficiency inactive-data-compression show -volume vol1_dst_dst -instance Volume: vol1_dst_dst Vserver: svm3 Is Enabled: true Scan Mode: - Progress: IDLE Status: SUCCESS Compression Algorithm: lzopro Failure Reason: - Total Blocks: - Total blocks Processed: - Percentage: - Phase1 L1s Processed: - Phase1 Lns Skipped: - Phase2 Total Blocks: - Phase2 Blocks Processed: - Number of Cold Blocks Encountered: 0 Number of Repacked Blocks: 0 Number of Compression Done Blocks: 0 Number of Vol-Overwrites: 0 Time since Last Inactive Data Compression Scan started(sec): 136 Time since Last Inactive Data Compression Scan ended(sec): 135 Time since Last Successful Inactive Data Compression Scan started(sec): - Time since Last Successful Inactive Data Compression Scan ended(sec): 135 Average time for Cold Data Compression(sec): 0 Tuning Enabled: true Threshold: 14 Threshold Upper Limit: 21 Threshold Lower Limit: 14 Client Read history window: 14 Incompressible Data Percentage: 0% ::*> volume show -volume vol1* -fields available, filesystem-size, total, used, percent-used, performance-tier-inactive-user-data, performance-tier-inactive-user-data-percent, size, dedupe-space-saved, dedupe-space-saved-percent, dedupe-space-shared, logical-used, logical-used-percent,logical-used-by-afs, logical-available vserver volume size available filesystem-size total used percent-used dedupe-space-saved dedupe-space-saved-percent dedupe-space-shared logical-used logical-used-percent logical-available logical-used-by-afs performance-tier-inactive-user-data performance-tier-inactive-user-data-percent ------- ------ ---- --------- --------------- ------- ------ ------------ ------------------ -------------------------- ------------------- ------------ -------------------- ----------------- ------------------- ----------------------------------- ------------------------------------------- svm vol1 16GB 9.89GB 16GB 15.20GB 5.31GB 34% 774.5MB 12% 774.5MB 6.06GB 40% - 6.06GB 0B 0% svm2 vol1_dst 6.37GB 967.6MB 6.37GB 6.05GB 5.11GB 84% 3.89GB 43% 2.09GB 8.98GB 148% - 6.04GB 0B 0% svm3 vol1_dst_dst 6.56GB 1.12GB 6.56GB 6.56GB 5.44GB 82% 3.83GB 41% 2.13GB 9.25GB 141% - 6.04GB 0B 0% 3 entries were displayed. ::*> volume show-footprint -volume vol1_dst_dst Vserver : svm3 Volume : vol1_dst_dst Feature Used Used% -------------------------------- ---------- ----- Volume Data Footprint 5.44GB 1% Footprint in Performance Tier 5.48GB 100% Footprint in FSxFabricpoolObjectStore 0B 0% Volume Guarantee 0B 0% Flexible Volume Metadata 44.31MB 0% Deduplication Metadata 12.30MB 0% Deduplication 12.30MB 0% Delayed Frees 40.04MB 0% File Operation Metadata 4KB 0% Total Footprint 5.53GB 1% Effective Total Footprint 5.53GB 1% ::*> aggr show-efficiency -instance Name of the Aggregate: aggr1 Node where Aggregate Resides: FsxId0ab6f9b00824a187c-01 Logical Size Used by Volumes, Clones, Snapshot Copies in the Aggregate: 63.60GB Total Physical Used: 10.76GB Total Storage Efficiency Ratio: 5.91:1 Total Data Reduction Logical Used Without Snapshots: 18.06GB Total Data Reduction Physical Used Without Snapshots: 6.58GB Total Data Reduction Efficiency Ratio Without Snapshots: 2.74:1 Total Data Reduction Logical Used without snapshots and flexclones: 18.06GB Total Data Reduction Physical Used without snapshots and flexclones: 6.58GB Total Data Reduction Efficiency Ratio without snapshots and flexclones: 2.74:1 Total Logical Size Used by Volumes, Clones, Snapshot Copies in the FabricPool Performance Tier: 63.73GB Total Physical Used in FabricPool Performance Tier: 10.96GB Total FabricPool Performance Tier Storage Efficiency Ratio: 5.81:1 Total Data Reduction Logical Used without snapshots and flexclones in the FabricPool Performance Tier: 18.19GB Total Data Reduction Physical Used without snapshots and flexclones in the FabricPool Performance Tier: 6.78GB Total FabricPool Performance Tier Data Reduction Efficiency Ratio without snapshots and flexclones: 2.68:1 Logical Space Used for All Volumes: 18.06GB Physical Space Used for All Volumes: 9.58GB Space Saved by Volume Deduplication: 8.48GB Space Saved by Volume Deduplication and pattern detection: 8.48GB Volume Deduplication Savings ratio: 1.89:1 Space Saved by Volume Compression: 0B Volume Compression Savings ratio: 1.00:1 Space Saved by Inline Zero Pattern Detection: 0B Volume Data Reduction SE Ratio: 1.89:1 Logical Space Used by the Aggregate: 16.66GB Physical Space Used by the Aggregate: 10.76GB Space Saved by Aggregate Data Reduction: 5.89GB Aggregate Data Reduction SE Ratio: 1.55:1 Logical Size Used by Snapshot Copies: 45.54GB Physical Size Used by Snapshot Copies: 6.46GB Snapshot Volume Data Reduction Ratio: 7.04:1 Logical Size Used by FlexClone Volumes: 0B Physical Sized Used by FlexClone Volumes: 0B FlexClone Volume Data Reduction Ratio: 1.00:1 Snapshot And FlexClone Volume Data Reduction SE Ratio: 7.04:1 Number of Volumes Offline: 0 Number of SIS Disabled Volumes: 3 Number of SIS Change Log Disabled Volumes: 0 ::*> snapshot show -volume vol1* ---Blocks--- Vserver Volume Snapshot Size Total% Used% -------- -------- ------------------------------------- -------- ------ ----- svm vol1 snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_015623 228KB 0% 0% test.2023-12-22_0533 160KB 0% 0% snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_053528 148KB 0% 0% svm2 vol1_dst snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_015623 1.91GB 30% 47% test.2023-12-22_0533 893.2MB 14% 29% snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_053528 480.5MB 7% 18% svm3 vol1_dst_dst snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_015623 1.91GB 29% 46% test.2023-12-22_0533 1.29GB 20% 37% snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_053528 148KB 0% 0% 9 entries were displayed.
vol1_dst_dst
の重複排除量が3.83GBとvol1_dst
とおおよそ同じ値になりました。
つまりはカスケードSnapMirrorで2つ目のボリュームで実行したStorage Efficiencyの結果を転送先ボリュームでも維持できることが分かります。
また、Space Saved by Aggregate Data Reduction
は4.96GBから5.89GBと1GB増加していました。
既にSnapMirrorで転送したファイルと重複している場合に重複排除が効くのか
svm のボリュームのTiering PolicyをAllに変更
あまりカスケードSnapMirrorと関係ありませんが、既にSnapMirrorで転送したファイルと重複している場合に重複排除が効くのかを確認してみます。
ボリューム内にあるファイルをコピーした際に重複排除が行われないようにvol1
のTiering PolicyをAllに変更します。
::*> volume modify -vserver svm -volume vol1 -tiering-policy all Volume modify successful on volume vol1 of Vserver svm. ::*> volume show-footprint -volume vol1 Vserver : svm Volume : vol1 Feature Used Used% -------------------------------- ---------- ----- Volume Data Footprint 5.31GB 1% Footprint in Performance Tier 2.63GB 48% Footprint in FSxFabricpoolObjectStore 2.82GB 52% Volume Guarantee 0B 0% Flexible Volume Metadata 92.66MB 0% Deduplication Metadata 6.02MB 0% Deduplication 6.02MB 0% Delayed Frees 144.8MB 0% File Operation Metadata 4KB 0% Total Footprint 5.54GB 1% Footprint Data Reduction in capacity tier 1.13GB - Effective Total Footprint 4.42GB 0% ::*> volume show-footprint -volume vol1 Vserver : svm Volume : vol1 Feature Used Used% -------------------------------- ---------- ----- Volume Data Footprint 5.33GB 1% Footprint in Performance Tier 234.9MB 4% Footprint in FSxFabricpoolObjectStore 5.24GB 96% Volume Guarantee 0B 0% Flexible Volume Metadata 92.66MB 0% Deduplication Metadata 6.02MB 0% Deduplication 6.02MB 0% Delayed Frees 146.0MB 0% File Operation Metadata 4KB 0% Total Footprint 5.57GB 1% Footprint Data Reduction in capacity tier 1.94GB - Effective Total Footprint 3.63GB 0%
96%のデータがキャパシティプールストレージにTieringされました。
aggregateの情報を確認します。
::*> aggr show-efficiency -instance Name of the Aggregate: aggr1 Node where Aggregate Resides: FsxId0ab6f9b00824a187c-01 Logical Size Used by Volumes, Clones, Snapshot Copies in the Aggregate: 63.60GB Total Physical Used: 10.25GB Total Storage Efficiency Ratio: 6.20:1 Total Data Reduction Logical Used Without Snapshots: 18.06GB Total Data Reduction Physical Used Without Snapshots: 6.13GB Total Data Reduction Efficiency Ratio Without Snapshots: 2.94:1 Total Data Reduction Logical Used without snapshots and flexclones: 18.06GB Total Data Reduction Physical Used without snapshots and flexclones: 6.13GB Total Data Reduction Efficiency Ratio without snapshots and flexclones: 2.94:1 Total Logical Size Used by Volumes, Clones, Snapshot Copies in the FabricPool Performance Tier: 43.42GB Total Physical Used in FabricPool Performance Tier: 7.23GB Total FabricPool Performance Tier Storage Efficiency Ratio: 6.00:1 Total Data Reduction Logical Used without snapshots and flexclones in the FabricPool Performance Tier: 12.38GB Total Data Reduction Physical Used without snapshots and flexclones in the FabricPool Performance Tier: 3.13GB Total FabricPool Performance Tier Data Reduction Efficiency Ratio without snapshots and flexclones: 3.96:1 Logical Space Used for All Volumes: 18.06GB Physical Space Used for All Volumes: 9.58GB Space Saved by Volume Deduplication: 8.48GB Space Saved by Volume Deduplication and pattern detection: 8.48GB Volume Deduplication Savings ratio: 1.89:1 Space Saved by Volume Compression: 0B Volume Compression Savings ratio: 1.00:1 Space Saved by Inline Zero Pattern Detection: 0B Volume Data Reduction SE Ratio: 1.89:1 Logical Space Used by the Aggregate: 16.15GB Physical Space Used by the Aggregate: 10.25GB Space Saved by Aggregate Data Reduction: 5.89GB Aggregate Data Reduction SE Ratio: 1.58:1 Logical Size Used by Snapshot Copies: 45.54GB Physical Size Used by Snapshot Copies: 6.49GB Snapshot Volume Data Reduction Ratio: 7.02:1 Logical Size Used by FlexClone Volumes: 0B Physical Sized Used by FlexClone Volumes: 0B FlexClone Volume Data Reduction Ratio: 1.00:1 Snapshot And FlexClone Volume Data Reduction SE Ratio: 7.02:1 Number of Volumes Offline: 0 Number of SIS Disabled Volumes: 3 Number of SIS Change Log Disabled Volumes: 0
Space Saved by Aggregate Data Reduction
はTiering前後で変わっていませんね。そのため、Tieringをしてもaggregateで実行されたデータ削減の効果が維持されていることが分かります。
この処理は以下のアップデートによるものです。文章だけ見ると「キャパシティプールストレージ上で追加の圧縮をかけられるようになった」ように認識してしまいそうですが、「キャパシティプールストレージに階層化するタイミングで圧縮するようになった」というものです。
テスト用ファイルの追加
作成するファイルは以下の3つです。
abcde
で埋めた1GiBのテキストファイル- 既にボリューム上に存在しているランダムブロックの1GiBのバイナリファイルのコピー
作成時のログは以下のとおりです。
$ sudo cp /mnt/fsxn/vol1/urandom_block_file /mnt/fsxn/vol1/urandom_block_file_copy2 $ yes abcde | tr -d '\n' | sudo dd of=/mnt/fsxn/vol1/abcde_padding_file bs=1024 count=1024K 1048576+0 records in 1048576+0 records out 1073741824 bytes (1.1 GB, 1.0 GiB) copied, 6.88529 s, 156 MB/s $ ls -lh /mnt/fsxn/vol1 total 8.1G -rw-r--r--. 1 root root 1.0G Dec 22 01:46 1_padding_file -rw-r--r--. 1 root root 1.0G Dec 22 05:28 a_padding_file -rw-r--r--. 1 root root 1.0G Dec 22 06:55 abcde_padding_file -rw-r--r--. 1 root root 1.0G Dec 22 01:47 urandom_block_file -rw-r--r--. 1 root root 1.0G Dec 22 05:02 urandom_block_file2 -rw-r--r--. 1 root root 1.0G Dec 22 05:02 urandom_block_file2_copy -rw-r--r--. 1 root root 1.0G Dec 22 01:47 urandom_block_file_copy -rw-r--r--. 1 root root 1.0G Dec 22 06:41 urandom_block_file_copy2 $ df -hT -t nfs4 Filesystem Type Size Used Avail Use% Mounted on svm-0058ae83d258ab2e3.fs-0ab6f9b00824a187c.fsx.us-east-1.amazonaws.com:/vol1 nfs4 16G 7.4G 7.9G 49% /mnt/fsxn/vol1
ボリューム、aggregateの情報を確認します。
::*> volume show -volume vol1* -fields available, filesystem-size, total, used, percent-used, performance-tier-inactive-user-data, performance-tier-inactive-user-data-percent, size, dedupe-space-saved, dedupe-space-saved-percent, dedupe-space-shared, logical-used, logical-used-percent,logical-used-by-afs, logical-available vserver volume size available filesystem-size total used percent-used dedupe-space-saved dedupe-space-saved-percent dedupe-space-shared logical-used logical-used-percent logical-available logical-used-by-afs performance-tier-inactive-user-data performance-tier-inactive-user-data-percent ------- ------ ---- --------- --------------- ------- ------ ------------ ------------------ -------------------------- ------------------- ------------ -------------------- ----------------- ------------------- ----------------------------------- ------------------------------------------- svm vol1 16GB 7.87GB 16GB 15.20GB 7.33GB 48% 774.5MB 9% 774.5MB 8.09GB 53% - 8.09GB - - svm2 vol1_dst 6.37GB 967.5MB 6.37GB 6.05GB 5.11GB 84% 3.89GB 43% 2.09GB 8.98GB 148% - 6.04GB 0B 0% svm3 vol1_dst_dst 6.56GB 1.12GB 6.56GB 6.56GB 5.44GB 82% 3.83GB 41% 2.13GB 9.25GB 141% - 6.04GB 0B 0% 3 entries were displayed. ::*> volume show-footprint -volume vol1 Vserver : svm Volume : vol1 Feature Used Used% -------------------------------- ---------- ----- Volume Data Footprint 7.36GB 1% Footprint in Performance Tier 1.26GB 17% Footprint in FSxFabricpoolObjectStore 6.25GB 83% Volume Guarantee 0B 0% Flexible Volume Metadata 92.66MB 0% Deduplication Metadata 16.79MB 0% Deduplication 8.78MB 0% Temporary Deduplication 8.02MB 0% Delayed Frees 155.7MB 0% File Operation Metadata 4KB 0% Total Footprint 7.62GB 1% Footprint Data Reduction in capacity tier 1.94GB - Effective Total Footprint 5.68GB 1% ::*> aggr show-efficiency -instance Name of the Aggregate: aggr1 Node where Aggregate Resides: FsxId0ab6f9b00824a187c-01 Logical Size Used by Volumes, Clones, Snapshot Copies in the Aggregate: 65.61GB Total Physical Used: 11.72GB Total Storage Efficiency Ratio: 5.60:1 Total Data Reduction Logical Used Without Snapshots: 20.07GB Total Data Reduction Physical Used Without Snapshots: 7.63GB Total Data Reduction Efficiency Ratio Without Snapshots: 2.63:1 Total Data Reduction Logical Used without snapshots and flexclones: 20.07GB Total Data Reduction Physical Used without snapshots and flexclones: 7.63GB Total Data Reduction Efficiency Ratio without snapshots and flexclones: 2.63:1 Total Logical Size Used by Volumes, Clones, Snapshot Copies in the FabricPool Performance Tier: 46.43GB Total Physical Used in FabricPool Performance Tier: 7.71GB Total FabricPool Performance Tier Storage Efficiency Ratio: 6.02:1 Total Data Reduction Logical Used without snapshots and flexclones in the FabricPool Performance Tier: 13.48GB Total Data Reduction Physical Used without snapshots and flexclones in the FabricPool Performance Tier: 3.63GB Total FabricPool Performance Tier Data Reduction Efficiency Ratio without snapshots and flexclones: 3.71:1 Logical Space Used for All Volumes: 20.07GB Physical Space Used for All Volumes: 11.58GB Space Saved by Volume Deduplication: 8.48GB Space Saved by Volume Deduplication and pattern detection: 8.48GB Volume Deduplication Savings ratio: 1.73:1 Space Saved by Volume Compression: 0B Volume Compression Savings ratio: 1.00:1 Space Saved by Inline Zero Pattern Detection: 0B Volume Data Reduction SE Ratio: 1.73:1 Logical Space Used by the Aggregate: 18.60GB Physical Space Used by the Aggregate: 11.72GB Space Saved by Aggregate Data Reduction: 6.88GB Aggregate Data Reduction SE Ratio: 1.59:1 Logical Size Used by Snapshot Copies: 45.54GB Physical Size Used by Snapshot Copies: 6.49GB Snapshot Volume Data Reduction Ratio: 7.02:1 Logical Size Used by FlexClone Volumes: 0B Physical Sized Used by FlexClone Volumes: 0B FlexClone Volume Data Reduction Ratio: 1.00:1 Snapshot And FlexClone Volume Data Reduction SE Ratio: 7.02:1 Number of Volumes Offline: 0 Number of SIS Disabled Volumes: 3 Number of SIS Change Log Disabled Volumes: 0
追加の重複排除は発生していないようです。
また、Space Saved by Aggregate Data Reduction
が1GB分増えています。これはabcde
で埋められた1GiBのテキストファイル分だと考えます。
svm と svm2 のボリューム間のSnapMirror update
svm と svm2 のボリューム間のSnapMirrorの差分転送を行います。
::*> snapmirror update -destination-path svm2:vol1_dst Operation is queued: snapmirror update of destination "svm2:vol1_dst". ::*> snapmirror show Progress Source Destination Mirror Relationship Total Last Path Type Path State Status Progress Healthy Updated ----------- ---- ------------ ------- -------------- --------- ------- -------- svm:vol1 XDP svm2:vol1_dst Snapmirrored Transferring 0B true 12/22 06:59:00 svm2:vol1_dst XDP svm3:vol1_dst_dst Snapmirrored Idle - true - 2 entries were displayed. ::*> snapmirror show Progress Source Destination Mirror Relationship Total Last Path Type Path State Status Progress Healthy Updated ----------- ---- ------------ ------- -------------- --------- ------- -------- svm:vol1 XDP svm2:vol1_dst Snapmirrored Transferring 119.2MB true 12/22 06:59:13 svm2:vol1_dst XDP svm3:vol1_dst_dst Snapmirrored Idle - true - 2 entries were displayed. ::*> snapmirror show Progress Source Destination Mirror Relationship Total Last Path Type Path State Status Progress Healthy Updated ----------- ---- ------------ ------- -------------- --------- ------- -------- svm:vol1 XDP svm2:vol1_dst Snapmirrored Transferring 436.0MB true 12/22 06:59:29 svm2:vol1_dst XDP svm3:vol1_dst_dst Snapmirrored Idle - true - 2 entries were displayed. ::*> snapmirror show Progress Source Destination Mirror Relationship Total Last Path Type Path State Status Progress Healthy Updated ----------- ---- ------------ ------- -------------- --------- ------- -------- svm:vol1 XDP svm2:vol1_dst Snapmirrored Transferring 1.12GB true 12/22 07:00:31 svm2:vol1_dst XDP svm3:vol1_dst_dst Snapmirrored Idle - true - 2 entries were displayed. ::*> snapmirror show Progress Source Destination Mirror Relationship Total Last Path Type Path State Status Progress Healthy Updated ----------- ---- ------------ ------- -------------- --------- ------- -------- svm:vol1 XDP svm2:vol1_dst Snapmirrored Transferring 1.12GB true 12/22 07:00:31 svm2:vol1_dst XDP svm3:vol1_dst_dst Snapmirrored Idle - true - 2 entries were displayed. ::*> snapmirror show Progress Source Destination Mirror Relationship Total Last Path Type Path State Status Progress Healthy Updated ----------- ---- ------------ ------- -------------- --------- ------- -------- svm:vol1 XDP svm2:vol1_dst Snapmirrored Transferring 1.37GB true 12/22 07:01:02 svm2:vol1_dst XDP svm3:vol1_dst_dst Snapmirrored Idle - true - 2 entries were displayed. ::*> snapmirror show Progress Source Destination Mirror Relationship Total Last Path Type Path State Status Progress Healthy Updated ----------- ---- ------------ ------- -------------- --------- ------- -------- svm:vol1 XDP svm2:vol1_dst Snapmirrored Transferring 1.49GB true 12/22 07:01:18 svm2:vol1_dst XDP svm3:vol1_dst_dst Snapmirrored Idle - true - 2 entries were displayed. ::*> snapmirror show Progress Source Destination Mirror Relationship Total Last Path Type Path State Status Progress Healthy Updated ----------- ---- ------------ ------- -------------- --------- ------- -------- svm:vol1 XDP svm2:vol1_dst Snapmirrored Transferring 1.75GB true 12/22 07:02:04 svm2:vol1_dst XDP svm3:vol1_dst_dst Snapmirrored Idle - true - 2 entries were displayed. ::*> volume efficiency show -volume vol1* -fields changelog-usage, changelog-size, logical-data-size, state, progress, last-op-size vserver volume state progress last-op-size changelog-usage changelog-size logical-data-size ------- ------ ------- -------------------- ------------ --------------- -------------- ----------------- svm vol1 Enabled 880112 KB (67%) Done 1.76GB 9% 32.43MB 8.09GB svm2 vol1_dst Enabled Idle for 01:13:40 9.22GB 0% 448KB 6.10GB svm3 vol1_dst_dst Enabled Idle for 01:05:43 5.24GB 0% 836KB 6.06GB 3 entries were displayed. ::*> ::*> volume efficiency show -volume vol1* -fields changelog-usage, changelog-size, logical-data-size, state, progress, last-op-size vserver volume state progress last-op-size changelog-usage changelog-size logical-data-size ------- ------ ------- -------------------- ------------ --------------- -------------- ----------------- svm vol1 Enabled 895604 KB (68%) Done 1.76GB 9% 32.43MB 8.09GB svm2 vol1_dst Enabled 830968 KB (39%) Done 9.22GB 0% 448KB 8.15GB svm3 vol1_dst_dst Enabled Idle for 01:06:17 5.24GB 0% 836KB 6.06GB 3 entries were displayed. ::*> snapmirror show Progress Source Destination Mirror Relationship Total Last Path Type Path State Status Progress Healthy Updated ----------- ---- ------------ ------- -------------- --------- ------- -------- svm:vol1 XDP svm2:vol1_dst Snapmirrored Idle - true - svm2:vol1_dst XDP svm3:vol1_dst_dst Snapmirrored Idle - true - 2 entries were displayed. ::*> snapmirror show -destination-path svm2:vol1_dst Source Path: svm:vol1 Source Cluster: - Source Vserver: svm Source Volume: vol1 Destination Path: svm2:vol1_dst Destination Cluster: - Destination Vserver: svm2 Destination Volume: vol1_dst Relationship Type: XDP Relationship Group Type: none Managing Vserver: svm2 SnapMirror Schedule: - SnapMirror Policy Type: async-mirror SnapMirror Policy: MirrorAllSnapshots Tries Limit: - Throttle (KB/sec): unlimited Consistency Group Item Mappings: - Current Transfer Throttle (KB/sec): - Mirror State: Snapmirrored Relationship Status: Idle File Restore File Count: - File Restore File List: - Transfer Snapshot: - Snapshot Progress: - Total Progress: - Network Compression Ratio: - Snapshot Checkpoint: - Newest Snapshot: snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_065900 Newest Snapshot Timestamp: 12/22 06:59:00 Exported Snapshot: snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_065900 Exported Snapshot Timestamp: 12/22 06:59:00 Healthy: true Relationship ID: 4f726f26-a06d-11ee-981e-bdd56ead09c8 Source Vserver UUID: 04ee1778-a058-11ee-981e-bdd56ead09c8 Destination Vserver UUID: 5af907bb-a065-11ee-981e-bdd56ead09c8 Current Operation ID: - Transfer Type: - Transfer Error: - Last Transfer Type: update Last Transfer Error: - Last Transfer Error Codes: - Last Transfer Size: 2.04GB Last Transfer Network Compression Ratio: 1:1 Last Transfer Duration: 0:3:59 Last Transfer From: svm:vol1 Last Transfer End Timestamp: 12/22 07:02:59 Unhealthy Reason: - Progress Last Updated: - Relationship Capability: 8.2 and above Lag Time: 0:4:29 Current Transfer Priority: - SMTape Operation: - Destination Volume Node Name: FsxId0ab6f9b00824a187c-01 Identity Preserve Vserver DR: - Number of Successful Updates: 3 Number of Failed Updates: 0 Number of Successful Resyncs: 0 Number of Failed Resyncs: 0 Number of Successful Breaks: 0 Number of Failed Breaks: 0 Total Transfer Bytes: 5853976980 Total Transfer Time in Seconds: 287 Source Volume MSIDs Preserved: - OpMask: ffffffffffffffff Is Auto Expand Enabled: - Percent Complete for Current Status: -
転送サイズは2.04GBでした。
また、SnapMirrorの転送中にStorage Efficiencyを確認すると、vol1_dst
でStorage Efficiencyが実行中でした。
Storage Efficiency、ボリューム、aggregate、Snapshotの情報を確認します。
::*> volume efficiency show -volume vol1* -fields changelog-usage, changelog-size, logical-data-size, state, progress, last-op-size vserver volume state progress last-op-size changelog-usage changelog-size logical-data-size ------- ------ ------- -------------------- ------------ --------------- -------------- ----------------- svm vol1 Enabled 922412 KB (70%) Done 1.76GB 9% 32.43MB 8.09GB svm2 vol1_dst Enabled Idle for 00:00:25 45.08MB 0% 700KB 8.10GB svm3 vol1_dst_dst Enabled Idle for 01:07:11 5.24GB 0% 836KB 6.06GB 3 entries were displayed. ::*> volume efficiency inactive-data-compression show -volume vol1_dst -instance Volume: vol1_dst Vserver: svm2 Is Enabled: true Scan Mode: - Progress: IDLE Status: SUCCESS Compression Algorithm: lzopro Failure Reason: - Total Blocks: - Total blocks Processed: - Percentage: - Phase1 L1s Processed: - Phase1 Lns Skipped: - Phase2 Total Blocks: - Phase2 Blocks Processed: - Number of Cold Blocks Encountered: 0 Number of Repacked Blocks: 0 Number of Compression Done Blocks: 0 Number of Vol-Overwrites: 0 Time since Last Inactive Data Compression Scan started(sec): 651 Time since Last Inactive Data Compression Scan ended(sec): 650 Time since Last Successful Inactive Data Compression Scan started(sec): - Time since Last Successful Inactive Data Compression Scan ended(sec): 650 Average time for Cold Data Compression(sec): 0 Tuning Enabled: true Threshold: 14 Threshold Upper Limit: 21 Threshold Lower Limit: 14 Client Read history window: 14 Incompressible Data Percentage: 61% ::*> volume show -volume vol1* -fields available, filesystem-size, total, used, percent-used, performance-tier-inactive-user-data, performance-tier-inactive-user-data-percent, size, dedupe-space-saved, dedupe-space-saved-percent, dedupe-space-shared, logical-used, logical-used-percent,logical-used-by-afs, logical-available vserver volume size available filesystem-size total used percent-used dedupe-space-saved dedupe-space-saved-percent dedupe-space-shared logical-used logical-used-percent logical-available logical-used-by-afs performance-tier-inactive-user-data performance-tier-inactive-user-data-percent ------- ------ ---- --------- --------------- ------- ------ ------------ ------------------ -------------------------- ------------------- ------------ -------------------- ----------------- ------------------- ----------------------------------- ------------------------------------------- svm vol1 16GB 7.86GB 16GB 15.20GB 7.33GB 48% 774.5MB 9% 774.5MB 8.09GB 53% - 8.09GB - - svm2 vol1_dst 7.54GB 1.14GB 7.54GB 7.16GB 6.02GB 84% 4.96GB 45% 3.03GB 10.95GB 153% - 8.08GB 0B 0% svm3 vol1_dst_dst 6.56GB 1.12GB 6.56GB 6.56GB 5.44GB 82% 3.83GB 41% 2.13GB 9.25GB 141% - 6.04GB 0B 0% 3 entries were displayed. ::*> volume show-footprint -volume vol1_dst Vserver : svm2 Volume : vol1_dst Feature Used Used% -------------------------------- ---------- ----- Volume Data Footprint 6.40GB 1% Footprint in Performance Tier 6.47GB 100% Footprint in FSxFabricpoolObjectStore 0B 0% Volume Guarantee 0B 0% Flexible Volume Metadata 55.75MB 0% Deduplication Metadata 12.30MB 0% Deduplication 12.30MB 0% Delayed Frees 70.04MB 0% File Operation Metadata 4KB 0% Total Footprint 6.54GB 1% Footprint Data Reduction 2.41GB 0% Auto Adaptive Compression 2.41GB 0% Effective Total Footprint 4.13GB 0% ::*> aggr show-efficiency -instance Name of the Aggregate: aggr1 Node where Aggregate Resides: FsxId0ab6f9b00824a187c-01 Logical Size Used by Volumes, Clones, Snapshot Copies in the Aggregate: 80.78GB Total Physical Used: 11.94GB Total Storage Efficiency Ratio: 6.77:1 Total Data Reduction Logical Used Without Snapshots: 22.08GB Total Data Reduction Physical Used Without Snapshots: 7.82GB Total Data Reduction Efficiency Ratio Without Snapshots: 2.82:1 Total Data Reduction Logical Used without snapshots and flexclones: 22.08GB Total Data Reduction Physical Used without snapshots and flexclones: 7.82GB Total Data Reduction Efficiency Ratio without snapshots and flexclones: 2.82:1 Total Logical Size Used by Volumes, Clones, Snapshot Copies in the FabricPool Performance Tier: 53.45GB Total Physical Used in FabricPool Performance Tier: 7.91GB Total FabricPool Performance Tier Storage Efficiency Ratio: 6.76:1 Total Data Reduction Logical Used without snapshots and flexclones in the FabricPool Performance Tier: 14.40GB Total Data Reduction Physical Used without snapshots and flexclones in the FabricPool Performance Tier: 3.81GB Total FabricPool Performance Tier Data Reduction Efficiency Ratio without snapshots and flexclones: 3.78:1 Logical Space Used for All Volumes: 22.08GB Physical Space Used for All Volumes: 12.53GB Space Saved by Volume Deduplication: 9.55GB Space Saved by Volume Deduplication and pattern detection: 9.55GB Volume Deduplication Savings ratio: 1.76:1 Space Saved by Volume Compression: 0B Volume Compression Savings ratio: 1.00:1 Space Saved by Inline Zero Pattern Detection: 0B Volume Data Reduction SE Ratio: 1.76:1 Logical Space Used by the Aggregate: 18.81GB Physical Space Used by the Aggregate: 11.94GB Space Saved by Aggregate Data Reduction: 6.88GB Aggregate Data Reduction SE Ratio: 1.58:1 Logical Size Used by Snapshot Copies: 58.70GB Physical Size Used by Snapshot Copies: 6.49GB Snapshot Volume Data Reduction Ratio: 9.04:1 Logical Size Used by FlexClone Volumes: 0B Physical Sized Used by FlexClone Volumes: 0B FlexClone Volume Data Reduction Ratio: 1.00:1 Snapshot And FlexClone Volume Data Reduction SE Ratio: 9.04:1 Number of Volumes Offline: 0 Number of SIS Disabled Volumes: 3 Number of SIS Change Log Disabled Volumes: 0 ::*> snapshot show -volume vol1* ---Blocks--- Vserver Volume Snapshot Size Total% Used% -------- -------- ------------------------------------- -------- ------ ----- svm vol1 test.2023-12-22_0533 160KB 0% 0% snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_053528 24.45MB 0% 0% snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_065900 164KB 0% 0% svm2 vol1_dst snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_015623 1.91GB 25% 38% test.2023-12-22_0533 893.2MB 12% 22% snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_053528 481.9MB 6% 13% snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_065900 148KB 0% 0% svm3 vol1_dst_dst snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_015623 1.91GB 29% 46% test.2023-12-22_0533 1.29GB 20% 37% snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_053528 148KB 0% 0% 10 entries were displayed.
重複排除量が3.89GBから4.96GBと1GB増加していました。
このことから既にSnapMirrorで転送したファイルと重複している場合に重複排除が効くことが分かります。
Space Saved by Aggregate Data Reduction
は変わりありませんでした。このことからキャパシティプールストレージ上のaggregateのデータ削減効果は、SnapMirrorで転送すると失われると考えられます。
vol1_dst で Inactive data compressionを実行
svm2のボリュームであるvol1_dst
でInactive data compressionを実行します。
::*> volume efficiency inactive-data-compression start -vserver svm2 -volume vol1_dst -inactive-days 0 Inactive data compression scan started on volume "vol1_dst" in Vserver "svm2" ::*> volume efficiency inactive-data-compression show -volume vol1_dst -instance Volume: vol1_dst Vserver: svm2 Is Enabled: true Scan Mode: default Progress: RUNNING Status: SUCCESS Compression Algorithm: lzopro Failure Reason: - Total Blocks: - Total blocks Processed: - Percentage: 57% Phase1 L1s Processed: 3609 Phase1 Lns Skipped: L1: 790 L2: 15 L3: 0 L4: 0 L5: 0 L6: 0 L7: 0 Phase2 Total Blocks: 2854496 Phase2 Blocks Processed: 1632009 Number of Cold Blocks Encountered: 248712 Number of Repacked Blocks: 0 Number of Compression Done Blocks: 235024 Number of Vol-Overwrites: 0 Time since Last Inactive Data Compression Scan started(sec): 5144 Time since Last Inactive Data Compression Scan ended(sec): 5124 Time since Last Successful Inactive Data Compression Scan started(sec): - Time since Last Successful Inactive Data Compression Scan ended(sec): 5124 Average time for Cold Data Compression(sec): 10 Tuning Enabled: true Threshold: 14 Threshold Upper Limit: 21 Threshold Lower Limit: 14 Client Read history window: 14 Incompressible Data Percentage: 61% ::*> volume efficiency inactive-data-compression show -volume vol1_dst -instance Volume: vol1_dst Vserver: svm2 Is Enabled: true Scan Mode: default Progress: RUNNING Status: SUCCESS Compression Algorithm: lzopro Failure Reason: - Total Blocks: - Total blocks Processed: - Percentage: 59% Phase1 L1s Processed: 3609 Phase1 Lns Skipped: L1: 790 L2: 15 L3: 0 L4: 0 L5: 0 L6: 0 L7: 0 Phase2 Total Blocks: 2854496 Phase2 Blocks Processed: 1684375 Number of Cold Blocks Encountered: 249440 Number of Repacked Blocks: 0 Number of Compression Done Blocks: 235040 Number of Vol-Overwrites: 0 Time since Last Inactive Data Compression Scan started(sec): 5157 Time since Last Inactive Data Compression Scan ended(sec): 5137 Time since Last Successful Inactive Data Compression Scan started(sec): - Time since Last Successful Inactive Data Compression Scan ended(sec): 5137 Average time for Cold Data Compression(sec): 10 Tuning Enabled: true Threshold: 14 Threshold Upper Limit: 21 Threshold Lower Limit: 14 Client Read history window: 14 Incompressible Data Percentage: 61% ::*> volume efficiency inactive-data-compression show -volume vol1_dst -instance Volume: vol1_dst Vserver: svm2 Is Enabled: true Scan Mode: default Progress: RUNNING Status: SUCCESS Compression Algorithm: lzopro Failure Reason: - Total Blocks: - Total blocks Processed: - Percentage: 69% Phase1 L1s Processed: 3609 Phase1 Lns Skipped: L1: 790 L2: 15 L3: 0 L4: 0 L5: 0 L6: 0 L7: 0 Phase2 Total Blocks: 2854496 Phase2 Blocks Processed: 1980084 Number of Cold Blocks Encountered: 252176 Number of Repacked Blocks: 0 Number of Compression Done Blocks: 235064 Number of Vol-Overwrites: 0 Time since Last Inactive Data Compression Scan started(sec): 5183 Time since Last Inactive Data Compression Scan ended(sec): 5163 Time since Last Successful Inactive Data Compression Scan started(sec): - Time since Last Successful Inactive Data Compression Scan ended(sec): 5163 Average time for Cold Data Compression(sec): 10 Tuning Enabled: true Threshold: 14 Threshold Upper Limit: 21 Threshold Lower Limit: 14 Client Read history window: 14 Incompressible Data Percentage: 61% ::*> volume efficiency inactive-data-compression show -volume vol1_dst -instance Volume: vol1_dst Vserver: svm2 Is Enabled: true Scan Mode: - Progress: IDLE Status: SUCCESS Compression Algorithm: lzopro Failure Reason: - Total Blocks: - Total blocks Processed: - Percentage: - Phase1 L1s Processed: - Phase1 Lns Skipped: - Phase2 Total Blocks: - Phase2 Blocks Processed: - Number of Cold Blocks Encountered: 257856 Number of Repacked Blocks: 0 Number of Compression Done Blocks: 235072 Number of Vol-Overwrites: 0 Time since Last Inactive Data Compression Scan started(sec): 81 Time since Last Inactive Data Compression Scan ended(sec): 20 Time since Last Successful Inactive Data Compression Scan started(sec): - Time since Last Successful Inactive Data Compression Scan ended(sec): 20 Average time for Cold Data Compression(sec): 27 Tuning Enabled: true Threshold: 14 Threshold Upper Limit: 21 Threshold Lower Limit: 14 Client Read history window: 14 Incompressible Data Percentage: 52%
235072ブロック追加で圧縮されたようです。
Storage Efficiency、ボリューム、aggregate、Snapshotの情報を確認します。
::*> volume efficiency show -volume vol1* -fields state, policy, storage-efficiency-mode, inline-compression, inline-dedupe, compression, data-compaction, auto-adaptive-compression-savings, auto-adaptive-compression-existing-volume,using-auto-adaptive-compression vserver volume state policy compression inline-compression storage-efficiency-mode inline-dedupe data-compaction auto-adaptive-compression-savings using-auto-adaptive-compression auto-adaptive-compression-existing-volume ------- ------ ------- ------ ----------- ------------------ ----------------------- ------------- --------------- --------------------------------- ------------------------------- ----------------------------------------- svm vol1 Enabled auto false true efficient false true true true false svm2 vol1_dst Enabled - false true efficient false true true true false svm3 vol1_dst_dst Enabled - false true efficient false true true true false 3 entries were displayed. ::*> volume show -volume vol1* -fields available, filesystem-size, total, used, percent-used, performance-tier-inactive-user-data, performance-tier-inactive-user-data-percent, size, dedupe-space-saved, dedupe-space-saved-percent, dedupe-space-shared, logical-used, logical-used-percent,logical-used-by-afs, logical-available vserver volume size available filesystem-size total used percent-used dedupe-space-saved dedupe-space-saved-percent dedupe-space-shared logical-used logical-used-percent logical-available logical-used-by-afs performance-tier-inactive-user-data performance-tier-inactive-user-data-percent ------- ------ ---- --------- --------------- ------- ------ ------------ ------------------ -------------------------- ------------------- ------------ -------------------- ----------------- ------------------- ----------------------------------- ------------------------------------------- svm vol1 16GB 7.86GB 16GB 15.20GB 7.33GB 48% 791.6MB 10% 791.6MB 8.11GB 53% - 8.11GB - - svm2 vol1_dst 7.54GB 1.14GB 7.54GB 7.16GB 6.02GB 84% 4.96GB 45% 3.03GB 10.95GB 153% - 8.08GB 0B 0% svm3 vol1_dst_dst 6.56GB 1.12GB 6.56GB 6.56GB 5.44GB 82% 3.83GB 41% 2.13GB 9.25GB 141% - 6.04GB 0B 0% 3 entries were displayed. ::*> volume show-footprint -volume vol1_dst Vserver : svm2 Volume : vol1_dst Feature Used Used% -------------------------------- ---------- ----- Volume Data Footprint 6.40GB 1% Footprint in Performance Tier 6.47GB 100% Footprint in FSxFabricpoolObjectStore 0B 0% Volume Guarantee 0B 0% Flexible Volume Metadata 55.75MB 0% Deduplication Metadata 12.30MB 0% Deduplication 12.30MB 0% Delayed Frees 71.98MB 0% File Operation Metadata 4KB 0% Total Footprint 6.54GB 1% Footprint Data Reduction 2.96GB 0% Auto Adaptive Compression 2.96GB 0% Effective Total Footprint 3.58GB 0% ::*> aggr show-efficiency -instance Name of the Aggregate: aggr1 Node where Aggregate Resides: FsxId0ab6f9b00824a187c-01 Logical Size Used by Volumes, Clones, Snapshot Copies in the Aggregate: 80.78GB Total Physical Used: 11.56GB Total Storage Efficiency Ratio: 6.99:1 Total Data Reduction Logical Used Without Snapshots: 22.08GB Total Data Reduction Physical Used Without Snapshots: 7.68GB Total Data Reduction Efficiency Ratio Without Snapshots: 2.88:1 Total Data Reduction Logical Used without snapshots and flexclones: 22.08GB Total Data Reduction Physical Used without snapshots and flexclones: 7.68GB Total Data Reduction Efficiency Ratio without snapshots and flexclones: 2.88:1 Total Logical Size Used by Volumes, Clones, Snapshot Copies in the FabricPool Performance Tier: 53.45GB Total Physical Used in FabricPool Performance Tier: 7.55GB Total FabricPool Performance Tier Storage Efficiency Ratio: 7.08:1 Total Data Reduction Logical Used without snapshots and flexclones in the FabricPool Performance Tier: 14.40GB Total Data Reduction Physical Used without snapshots and flexclones in the FabricPool Performance Tier: 3.68GB Total FabricPool Performance Tier Data Reduction Efficiency Ratio without snapshots and flexclones: 3.91:1 Logical Space Used for All Volumes: 22.08GB Physical Space Used for All Volumes: 12.52GB Space Saved by Volume Deduplication: 9.56GB Space Saved by Volume Deduplication and pattern detection: 9.56GB Volume Deduplication Savings ratio: 1.76:1 Space Saved by Volume Compression: 0B Volume Compression Savings ratio: 1.00:1 Space Saved by Inline Zero Pattern Detection: 0B Volume Data Reduction SE Ratio: 1.76:1 Logical Space Used by the Aggregate: 19.33GB Physical Space Used by the Aggregate: 11.56GB Space Saved by Aggregate Data Reduction: 7.77GB Aggregate Data Reduction SE Ratio: 1.67:1 Logical Size Used by Snapshot Copies: 58.70GB Physical Size Used by Snapshot Copies: 6.49GB Snapshot Volume Data Reduction Ratio: 9.04:1 Logical Size Used by FlexClone Volumes: 0B Physical Sized Used by FlexClone Volumes: 0B FlexClone Volume Data Reduction Ratio: 1.00:1 Snapshot And FlexClone Volume Data Reduction SE Ratio: 9.04:1 Number of Volumes Offline: 0 Number of SIS Disabled Volumes: 3 Number of SIS Change Log Disabled Volumes: 0 ::*> snapshot show -volume vol1* ---Blocks--- Vserver Volume Snapshot Size Total% Used% -------- -------- ------------------------------------- -------- ------ ----- svm vol1 test.2023-12-22_0533 160KB 0% 0% snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_053528 24.45MB 0% 0% snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_065900 272KB 0% 0% svm2 vol1_dst snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_015623 1.91GB 25% 38% test.2023-12-22_0533 893.2MB 12% 22% snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_053528 481.9MB 6% 13% snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_065900 148KB 0% 0% svm3 vol1_dst_dst snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_015623 1.91GB 29% 46% test.2023-12-22_0533 1.29GB 20% 37% snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_053528 148KB 0% 0% 10 entries were displayed.
Space Saved by Aggregate Data Reduction
が6.88GBから7.77GBと900MB圧縮されています。追加したテキストファイル分が圧縮されたのだと考えます。
vol1_dst で Storage Efficiencyを実行
次にsvm2のボリュームであるvol1_dst
でStorage Efficiencyを実行します。
::*> volume efficiency start -vserver svm2 -volume vol1_dst -scan-old-data Warning: This operation scans all of the data in volume "vol1_dst" of Vserver "svm2". It might take a significant time, and degrade performance during that time. Do you want to continue? {y|n}: y The efficiency operation for volume "vol1_dst" of Vserver "svm2" has started. ::*> volume efficiency show -volume vol1* -fields changelog-usage, changelog-size, logical-data-size, state, progress, last-op-size vserver volume state progress last-op-size changelog-usage changelog-size logical-data-size ------- ------ ------- ----------------- ------------ --------------- -------------- ----------------- svm vol1 Enabled Idle for 00:08:46 1.70GB 9% 15.41MB 8.11GB svm2 vol1_dst Enabled 5846532 KB Scanned 45.08MB 0% 0B 8.07GB svm3 vol1_dst_dst Enabled Idle for 01:28:42 5.24GB 0% 836KB 6.06GB 3 entries were displayed. ::*> volume efficiency show -volume vol1* -fields changelog-usage, changelog-size, logical-data-size, state, progress, last-op-size vserver volume state progress last-op-size changelog-usage changelog-size logical-data-size ------- ------ ------- ----------------- ------------ --------------- -------------- ----------------- svm vol1 Enabled Idle for 00:12:00 1.70GB 9% 15.41MB 8.11GB svm2 vol1_dst Enabled 10234036 KB (95%) Done 45.08MB 0% 540KB 8.07GB svm3 vol1_dst_dst Enabled Idle for 01:31:56 5.24GB 0% 836KB 6.06GB 3 entries were displayed. ::*> volume efficiency show -volume vol1* -fields changelog-usage, changelog-size, logical-data-size, state, progress, last-op-size vserver volume state progress last-op-size changelog-usage changelog-size logical-data-size ------- ------ ------- ----------------- ------------ --------------- -------------- ----------------- svm vol1 Enabled Idle for 00:12:50 1.70GB 9% 15.41MB 8.11GB svm2 vol1_dst Enabled Idle for 00:00:42 12.21GB 0% 540KB 8.08GB svm3 vol1_dst_dst Enabled Idle for 01:32:46 5.24GB 0% 836KB 6.06GB 3 entries were displayed.
ボリューム、aggregateの情報を確認します。
::*> volume show -volume vol1* -fields available, filesystem-size, total, used, percent-used, performance-tier-inactive-user-data, performance-tier-inactive-user-data-percent, size, dedupe-space-saved, dedupe-space-saved-percent, dedupe-space-shared, logical-used, logical-used-percent,logical-used-by-afs, logical-available vserver volume size available filesystem-size total used percent-used dedupe-space-saved dedupe-space-saved-percent dedupe-space-shared logical-used logical-used-percent logical-available logical-used-by-afs performance-tier-inactive-user-data performance-tier-inactive-user-data-percent ------- ------ ---- --------- --------------- ------- ------ ------------ ------------------ -------------------------- ------------------- ------------ -------------------- ----------------- ------------------- ----------------------------------- ------------------------------------------- svm vol1 16GB 7.86GB 16GB 15.20GB 7.33GB 48% 791.6MB 10% 791.6MB 8.11GB 53% - 8.11GB - - svm2 vol1_dst 7.54GB 1.10GB 7.54GB 7.16GB 6.06GB 84% 4.97GB 45% 3.01GB 11.01GB 154% - 8.05GB 0B 0% svm3 vol1_dst_dst 6.56GB 1.12GB 6.56GB 6.56GB 5.44GB 82% 3.83GB 41% 2.13GB 9.25GB 141% - 6.04GB 0B 0% 3 entries were displayed. ::*> volume show-footprint -volume vol1_dst Vserver : svm2 Volume : vol1_dst Feature Used Used% -------------------------------- ---------- ----- Volume Data Footprint 6.44GB 1% Footprint in Performance Tier 6.51GB 100% Footprint in FSxFabricpoolObjectStore 0B 0% Volume Guarantee 0B 0% Flexible Volume Metadata 55.75MB 0% Deduplication Metadata 12.04MB 0% Deduplication 12.04MB 0% Delayed Frees 72.96MB 0% File Operation Metadata 4KB 0% Total Footprint 6.58GB 1% Footprint Data Reduction 2.98GB 0% Auto Adaptive Compression 2.98GB 0% Effective Total Footprint 3.60GB 0% ::*> aggr show-efficiency -instance Name of the Aggregate: aggr1 Node where Aggregate Resides: FsxId0ab6f9b00824a187c-01 Logical Size Used by Volumes, Clones, Snapshot Copies in the Aggregate: 80.78GB Total Physical Used: 11.90GB Total Storage Efficiency Ratio: 6.79:1 Total Data Reduction Logical Used Without Snapshots: 22.08GB Total Data Reduction Physical Used Without Snapshots: 7.93GB Total Data Reduction Efficiency Ratio Without Snapshots: 2.79:1 Total Data Reduction Logical Used without snapshots and flexclones: 22.08GB Total Data Reduction Physical Used without snapshots and flexclones: 7.93GB Total Data Reduction Efficiency Ratio without snapshots and flexclones: 2.79:1 Total Logical Size Used by Volumes, Clones, Snapshot Copies in the FabricPool Performance Tier: 53.43GB Total Physical Used in FabricPool Performance Tier: 7.87GB Total FabricPool Performance Tier Storage Efficiency Ratio: 6.79:1 Total Data Reduction Logical Used without snapshots and flexclones in the FabricPool Performance Tier: 14.38GB Total Data Reduction Physical Used without snapshots and flexclones in the FabricPool Performance Tier: 3.91GB Total FabricPool Performance Tier Data Reduction Efficiency Ratio without snapshots and flexclones: 3.68:1 Logical Space Used for All Volumes: 22.08GB Physical Space Used for All Volumes: 12.50GB Space Saved by Volume Deduplication: 9.58GB Space Saved by Volume Deduplication and pattern detection: 9.58GB Volume Deduplication Savings ratio: 1.77:1 Space Saved by Volume Compression: 0B Volume Compression Savings ratio: 1.00:1 Space Saved by Inline Zero Pattern Detection: 0B Volume Data Reduction SE Ratio: 1.77:1 Logical Space Used by the Aggregate: 19.67GB Physical Space Used by the Aggregate: 11.90GB Space Saved by Aggregate Data Reduction: 7.77GB Aggregate Data Reduction SE Ratio: 1.65:1 Logical Size Used by Snapshot Copies: 58.70GB Physical Size Used by Snapshot Copies: 6.57GB Snapshot Volume Data Reduction Ratio: 8.94:1 Logical Size Used by FlexClone Volumes: 0B Physical Sized Used by FlexClone Volumes: 0B FlexClone Volume Data Reduction Ratio: 1.00:1 Snapshot And FlexClone Volume Data Reduction SE Ratio: 8.94:1 Number of Volumes Offline: 0 Number of SIS Disabled Volumes: 3 Number of SIS Change Log Disabled Volumes: 0 ::*> snapshot show -volume vol1* ---Blocks--- Vserver Volume Snapshot Size Total% Used% -------- -------- ------------------------------------- -------- ------ ----- svm vol1 test.2023-12-22_0533 160KB 0% 0% snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_053528 24.45MB 0% 0% snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_065900 272KB 0% 0% svm2 vol1_dst snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_015623 1.91GB 25% 38% test.2023-12-22_0533 893.2MB 12% 22% snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_053528 481.9MB 6% 13% snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_065900 79.82MB 1% 2% svm3 vol1_dst_dst snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_015623 1.91GB 29% 46% test.2023-12-22_0533 1.29GB 20% 37% snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_053528 148KB 0% 0% 10 entries were displayed.
重複排除量は3.89GBから4.97GBと1GBほど増加していました。
svm2 と svm3 のボリューム間のSnapMirror update
svm2 と svm3 のボリューム間のSnapMirrorの差分転送を行います。
::*> snapmirror update -destination-path svm3:vol1_dst_dst Operation is queued: snapmirror update of destination "svm3:vol1_dst_dst". ::*> snapmirror show Progress Source Destination Mirror Relationship Total Last Path Type Path State Status Progress Healthy Updated ----------- ---- ------------ ------- -------------- --------- ------- -------- svm:vol1 XDP svm2:vol1_dst Snapmirrored Idle - true - svm2:vol1_dst XDP svm3:vol1_dst_dst Snapmirrored Finalizing 0B true 12/22 07:31:30 2 entries were displayed. ::*> snapmirror show Progress Source Destination Mirror Relationship Total Last Path Type Path State Status Progress Healthy Updated ----------- ---- ------------ ------- -------------- --------- ------- -------- svm:vol1 XDP svm2:vol1_dst Snapmirrored Idle - true - svm2:vol1_dst XDP svm3:vol1_dst_dst Snapmirrored Finalizing 389.6MB true 12/22 07:31:42 2 entries were displayed. ::*> snapmirror show Progress Source Destination Mirror Relationship Total Last Path Type Path State Status Progress Healthy Updated ----------- ---- ------------ ------- -------------- --------- ------- -------- svm:vol1 XDP svm2:vol1_dst Snapmirrored Idle - true - svm2:vol1_dst XDP svm3:vol1_dst_dst Snapmirrored Idle - true - 2 entries were displayed. ::*> snapmirror show -destination-path svm3:vol1_dst_dst Source Path: svm2:vol1_dst Source Cluster: - Source Vserver: svm2 Source Volume: vol1_dst Destination Path: svm3:vol1_dst_dst Destination Cluster: - Destination Vserver: svm3 Destination Volume: vol1_dst_dst Relationship Type: XDP Relationship Group Type: none Managing Vserver: svm3 SnapMirror Schedule: - SnapMirror Policy Type: async-mirror SnapMirror Policy: MirrorAllSnapshots Tries Limit: - Throttle (KB/sec): unlimited Consistency Group Item Mappings: - Current Transfer Throttle (KB/sec): - Mirror State: Snapmirrored Relationship Status: Idle File Restore File Count: - File Restore File List: - Transfer Snapshot: - Snapshot Progress: - Total Progress: - Network Compression Ratio: - Snapshot Checkpoint: - Newest Snapshot: snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_065900 Newest Snapshot Timestamp: 12/22 06:59:00 Exported Snapshot: snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_065900 Exported Snapshot Timestamp: 12/22 06:59:00 Healthy: true Relationship ID: b0b2694d-a084-11ee-981e-bdd56ead09c8 Source Vserver UUID: 5af907bb-a065-11ee-981e-bdd56ead09c8 Destination Vserver UUID: 674509fa-a065-11ee-981e-bdd56ead09c8 Current Operation ID: - Transfer Type: - Transfer Error: - Last Transfer Type: update Last Transfer Error: - Last Transfer Error Codes: - Last Transfer Size: 389.6MB Last Transfer Network Compression Ratio: 1:1 Last Transfer Duration: 0:0:29 Last Transfer From: svm2:vol1_dst Last Transfer End Timestamp: 12/22 07:31:59 Unhealthy Reason: - Progress Last Updated: - Relationship Capability: 8.2 and above Lag Time: 0:33:19 Current Transfer Priority: - SMTape Operation: - Destination Volume Node Name: FsxId0ab6f9b00824a187c-01 Identity Preserve Vserver DR: - Number of Successful Updates: 3 Number of Failed Updates: 0 Number of Successful Resyncs: 0 Number of Failed Resyncs: 0 Number of Successful Breaks: 0 Number of Failed Breaks: 0 Total Transfer Bytes: 4129216624 Total Transfer Time in Seconds: 107 Source Volume MSIDs Preserved: - OpMask: ffffffffffffffff Is Auto Expand Enabled: - Percent Complete for Current Status: -
転送量は389.6MBでした。
Storage Efficiency、ボリューム、aggregate、Snapshotの情報を確認します。
::*> volume efficiency show -volume vol1* -fields changelog-usage, changelog-size, logical-data-size, state, progress, last-op-size vserver volume state progress last-op-size changelog-usage changelog-size logical-data-size ------- ------ ------- ----------------- ------------ --------------- -------------- ----------------- svm vol1 Enabled Idle for 00:15:44 1.70GB 9% 15.41MB 8.11GB svm2 vol1_dst Enabled Idle for 00:03:36 12.21GB 0% 540KB 8.08GB svm3 vol1_dst_dst Enabled Idle for 00:00:17 83.62MB 0% 900KB 8.09GB 3 entries were displayed. ::*> volume efficiency inactive-data-compression show -volume vol1_dst_dst -instance Volume: vol1_dst_dst Vserver: svm3 Is Enabled: true Scan Mode: - Progress: IDLE Status: SUCCESS Compression Algorithm: lzopro Failure Reason: - Total Blocks: - Total blocks Processed: - Percentage: - Phase1 L1s Processed: - Phase1 Lns Skipped: - Phase2 Total Blocks: - Phase2 Blocks Processed: - Number of Cold Blocks Encountered: 0 Number of Repacked Blocks: 0 Number of Compression Done Blocks: 0 Number of Vol-Overwrites: 0 Time since Last Inactive Data Compression Scan started(sec): 18 Time since Last Inactive Data Compression Scan ended(sec): 17 Time since Last Successful Inactive Data Compression Scan started(sec): - Time since Last Successful Inactive Data Compression Scan ended(sec): 17 Average time for Cold Data Compression(sec): 0 Tuning Enabled: true Threshold: 14 Threshold Upper Limit: 21 Threshold Lower Limit: 14 Client Read history window: 14 Incompressible Data Percentage: 0% ::*> volume show -volume vol1* -fields available, filesystem-size, total, used, percent-used, performance-tier-inactive-user-data, performance-tier-inactive-user-data-percent, size, dedupe-space-saved, dedupe-space-saved-percent, dedupe-space-shared, logical-used, logical-used-percent,logical-used-by-afs, logical-available vserver volume size available filesystem-size total used percent-used dedupe-space-saved dedupe-space-saved-percent dedupe-space-shared logical-used logical-used-percent logical-available logical-used-by-afs performance-tier-inactive-user-data performance-tier-inactive-user-data-percent ------- ------ ---- --------- --------------- ------- ------ ------------ ------------------ -------------------------- ------------------- ------------ -------------------- ----------------- ------------------- ----------------------------------- ------------------------------------------- svm vol1 16GB 7.86GB 16GB 15.20GB 7.33GB 48% 791.6MB 10% 791.6MB 8.11GB 53% - 8.11GB - - svm2 vol1_dst 7.54GB 1.10GB 7.54GB 7.16GB 6.06GB 84% 4.97GB 45% 3.01GB 11.01GB 154% - 8.05GB 0B 0% svm3 vol1_dst_dst 7.67GB 1.35GB 7.67GB 7.67GB 6.32GB 82% 5.08GB 45% 2.88GB 11.37GB 148% - 8.06GB 0B 0% 3 entries were displayed. ::*> volume show-footprint -volume vol1_dst_dst Vserver : svm3 Volume : vol1_dst_dst Feature Used Used% -------------------------------- ---------- ----- Volume Data Footprint 6.32GB 1% Footprint in Performance Tier 6.40GB 100% Footprint in FSxFabricpoolObjectStore 0B 0% Volume Guarantee 0B 0% Flexible Volume Metadata 50.02MB 0% Deduplication Metadata 13.21MB 0% Deduplication 13.21MB 0% Delayed Frees 81.63MB 0% File Operation Metadata 4KB 0% Total Footprint 6.46GB 1% Effective Total Footprint 6.46GB 1% ::*> aggr show-efficiency -instance Name of the Aggregate: aggr1 Node where Aggregate Resides: FsxId0ab6f9b00824a187c-01 Logical Size Used by Volumes, Clones, Snapshot Copies in the Aggregate: 90.85GB Total Physical Used: 12.10GB Total Storage Efficiency Ratio: 7.51:1 Total Data Reduction Logical Used Without Snapshots: 24.10GB Total Data Reduction Physical Used Without Snapshots: 8.15GB Total Data Reduction Efficiency Ratio Without Snapshots: 2.96:1 Total Data Reduction Logical Used without snapshots and flexclones: 24.10GB Total Data Reduction Physical Used without snapshots and flexclones: 8.15GB Total Data Reduction Efficiency Ratio without snapshots and flexclones: 2.96:1 Total Logical Size Used by Volumes, Clones, Snapshot Copies in the FabricPool Performance Tier: 63.52GB Total Physical Used in FabricPool Performance Tier: 8.08GB Total FabricPool Performance Tier Storage Efficiency Ratio: 7.86:1 Total Data Reduction Logical Used without snapshots and flexclones in the FabricPool Performance Tier: 16.40GB Total Data Reduction Physical Used without snapshots and flexclones in the FabricPool Performance Tier: 4.15GB Total FabricPool Performance Tier Data Reduction Efficiency Ratio without snapshots and flexclones: 3.95:1 Logical Space Used for All Volumes: 24.10GB Physical Space Used for All Volumes: 13.27GB Space Saved by Volume Deduplication: 10.82GB Space Saved by Volume Deduplication and pattern detection: 10.82GB Volume Deduplication Savings ratio: 1.82:1 Space Saved by Volume Compression: 0B Volume Compression Savings ratio: 1.00:1 Space Saved by Inline Zero Pattern Detection: 0B Volume Data Reduction SE Ratio: 1.82:1 Logical Space Used by the Aggregate: 20.45GB Physical Space Used by the Aggregate: 12.10GB Space Saved by Aggregate Data Reduction: 8.36GB Aggregate Data Reduction SE Ratio: 1.69:1 Logical Size Used by Snapshot Copies: 66.76GB Physical Size Used by Snapshot Copies: 6.67GB Snapshot Volume Data Reduction Ratio: 10.01:1 Logical Size Used by FlexClone Volumes: 0B Physical Sized Used by FlexClone Volumes: 0B FlexClone Volume Data Reduction Ratio: 1.00:1 Snapshot And FlexClone Volume Data Reduction SE Ratio: 10.01:1 Number of Volumes Offline: 0 Number of SIS Disabled Volumes: 3 Number of SIS Change Log Disabled Volumes: 0 ::*> snapshot show -volume vol1* ---Blocks--- Vserver Volume Snapshot Size Total% Used% -------- -------- ------------------------------------- -------- ------ ----- svm vol1 test.2023-12-22_0533 160KB 0% 0% snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_053528 24.45MB 0% 0% snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_065900 272KB 0% 0% svm2 vol1_dst snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_015623 1.91GB 25% 38% test.2023-12-22_0533 893.2MB 12% 22% snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_053528 481.9MB 6% 13% snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_065900 79.82MB 1% 2% svm3 vol1_dst_dst snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_015623 1.91GB 25% 39% test.2023-12-22_0533 1.29GB 17% 30% snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_053528 101.9MB 1% 3% snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_065900 144KB 0% 0% 11 entries were displayed.
重複排除量が5.08GBとなっています。問題なく重複排除量が維持されていますね。
加えてvol1_dst_dst
のStorage Efficiencyのprogressが更新されていることからSnapMirrorで転送されたデータに対して重複排除が行われたことが分かります。
また、Space Saved by Aggregate Data Reduction
は7.77GBから8.36GBと600MBほど増加しています。
svm と svm3 のボリューム間のSnapMirror
SVMピアリングの追加
次にsvm と svm3 のボリューム間のSnapMirrorを試します。
以下NetApp公式ドキュメントに記載の通り、カスケードSnapMirrorの中間のボリュームが使用できなくなった場合、直接転送元ボリュームから最終的な転送先ボリューム間で同期を行うことができるようです。
Bのボリュームが使用できなくなった場合は、CとAの間の関係を同期することでAの保護を継続できます。ベースライン転送を新たに実行する必要はありません。再同期処理が終わると、AはBを迂回してCと直接ミラー関係を持つことになります。ただし、再同期処理を実行するときは、再同期によってSnapshotコピーが削除され、カスケード内の関係の共通Snapshotコピーが失われる可能性があることに注意してください。その場合、関係には新しいベースラインが必要になります。
次の図に、ミラー-ミラー カスケード チェーンを示します。
実際に試しみましょう。
まずは、svmとsvm3間のSVMピアリングを追加します。
::*> vserver peer create -vserver svm -peer-vserver svm3 -applications snapmirror Info: 'vserver peer create' command is successful. ::*> vserver peer show Peer Peer Peering Remote Vserver Vserver State Peer Cluster Applications Vserver ----------- ----------- ------------ ----------------- -------------- --------- svm svm2 peered FsxId0ab6f9b00824a187c snapmirror svm2 svm svm3 peered FsxId0ab6f9b00824a187c snapmirror svm3 svm2 svm peered FsxId0ab6f9b00824a187c snapmirror svm svm2 svm3 peered FsxId0ab6f9b00824a187c snapmirror svm3 svm3 svm peered FsxId0ab6f9b00824a187c snapmirror svm svm3 svm2 peered FsxId0ab6f9b00824a187c snapmirror svm2 6 entries were displayed.
svm2 と svm3 のボリューム間のSnapMirror relasionshipの削除
svmとsvm3のボリューム間のSnapMirror relasionshipを作成するにあたって、svm2 と svm3 のボリューム間のSnapMirror relasionshipの削除を行います。
これはONTAPの仕様上、SnapMirror relationshpの転送先ボリュームを重複する(fan-in)ことはできないためです。SanpMirror relasionshipが存在する場合、状態がBroken-offであっても以下のようにエラーになります。
::*> snapmirror create -source-path svm:vol1 -destination-path svm3:vol1_dst_dst -policy MirrorAllSnapshots Error: command failed: Relationship with destination svm3:vol1_dst_dst already exists.
SnapMirrorをカットオーバーします。
::*> snapmirror quiesce -destination-path svm3:vol1_dst_dst Operation succeeded: snapmirror quiesce for destination "svm3:vol1_dst_dst". ::*> snapmirror break -destination-path svm3:vol1_dst_dst Operation succeeded: snapmirror break for destination "svm3:vol1_dst_dst". ::*> snapmirror show Progress Source Destination Mirror Relationship Total Last Path Type Path State Status Progress Healthy Updated ----------- ---- ------------ ------- -------------- --------- ------- -------- svm:vol1 XDP svm2:vol1_dst Snapmirrored Idle - true - svm2:vol1_dst XDP svm3:vol1_dst_dst Broken-off Idle - true - 2 entries were displayed.
そして、SnapMirror relasionshipを削除します。
::*> snapmirror delete -destination-path svm3:vol1_dst_dst Operation succeeded: snapmirror delete for the relationship with destination "svm3:vol1_dst_dst". ::*> snapmirror show Progress Source Destination Mirror Relationship Total Last Path Type Path State Status Progress Healthy Updated ----------- ---- ------------ ------- -------------- --------- ------- -------- svm:vol1 XDP svm2:vol1_dst Snapmirrored Idle - true -
svm と svm3 のボリューム間のSnapMirrorの再同期
svm と svm3 のボリューム間のSnapMirrorの再同期をします。
まず、SnapMirror relationshipを作成します。
::*> snapmirror create -source-path svm:vol1 -destination-path svm3:vol1_dst_dst -policy MirrorAllSnapshots Operation succeeded: snapmirror create for the relationship with destination "svm3:vol1_dst_dst". ::*> snapmirror show Progress Source Destination Mirror Relationship Total Last Path Type Path State Status Progress Healthy Updated ----------- ---- ------------ ------- -------------- --------- ------- -------- svm:vol1 XDP svm2:vol1_dst Snapmirrored Idle - true - svm3:vol1_dst_dst Broken-off Idle - true - 2 entries were displayed.
SnapMirrorの差分同期を行います。
::*> snapmirror update -destination-path svm3:vol1_dst_dst Operation is queued: snapmirror update of destination "svm3:vol1_dst_dst". ::*> snapmirror show Progress Source Destination Mirror Relationship Total Last Path Type Path State Status Progress Healthy Updated ----------- ---- ------------ ------- -------------- --------- ------- -------- svm:vol1 XDP svm2:vol1_dst Snapmirrored Idle - true - svm3:vol1_dst_dst Broken-off Idle - false - 2 entries were displayed. ::*> snapmirror show -destination-path svm3:vol1_dst_dst Source Path: svm:vol1 Source Cluster: - Source Vserver: svm Source Volume: vol1 Destination Path: svm3:vol1_dst_dst Destination Cluster: - Destination Vserver: svm3 Destination Volume: vol1_dst_dst Relationship Type: XDP Relationship Group Type: none Managing Vserver: svm3 SnapMirror Schedule: - SnapMirror Policy Type: async-mirror SnapMirror Policy: MirrorAllSnapshots Tries Limit: - Throttle (KB/sec): unlimited Consistency Group Item Mappings: - Current Transfer Throttle (KB/sec): - Mirror State: Broken-off Relationship Status: Idle File Restore File Count: - File Restore File List: - Transfer Snapshot: - Snapshot Progress: - Total Progress: - Network Compression Ratio: - Snapshot Checkpoint: - Newest Snapshot: snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_065900 Newest Snapshot Timestamp: 12/22 06:59:00 Exported Snapshot: - Exported Snapshot Timestamp: - Healthy: false Relationship ID: 27f77b8a-a09e-11ee-981e-bdd56ead09c8 Source Vserver UUID: 04ee1778-a058-11ee-981e-bdd56ead09c8 Destination Vserver UUID: 674509fa-a065-11ee-981e-bdd56ead09c8 Current Operation ID: - Transfer Type: - Transfer Error: - Last Transfer Type: update Last Transfer Error: Destination svm3:vol1_dst_dst must be a data-protection volume. Last Transfer Error Codes: 6619546 Last Transfer Size: - Last Transfer Network Compression Ratio: - Last Transfer Duration: - Last Transfer From: svm:vol1 Last Transfer End Timestamp: 12/22 07:46:29 Unhealthy Reason: Transfer failed. Progress Last Updated: - Relationship Capability: 8.2 and above Lag Time: - Current Transfer Priority: - SMTape Operation: - Destination Volume Node Name: FsxId0ab6f9b00824a187c-01 Identity Preserve Vserver DR: - Number of Successful Updates: 0 Number of Failed Updates: 1 Number of Successful Resyncs: 0 Number of Failed Resyncs: 0 Number of Successful Breaks: 0 Number of Failed Breaks: 0 Total Transfer Bytes: 0 Total Transfer Time in Seconds: 0 Source Volume MSIDs Preserved: - OpMask: ffffffffffffffff Is Auto Expand Enabled: - Percent Complete for Current Status: -
svm3のvol1_dst_dst
のボリュームタイプがDP
ではないとのことで失敗しました。
以下のKBにはsnapmirror resyncを実行してSnapMirror relationshipを再確立するようにと記載がありました。
- Run a snapmirror resync to re-establish the SnapMirror relationship
- Re-run the update and verify completion status
SnapMirror update fails with error "must be a data-protection volume" - NetApp Knowledge Base
実際に試します。
::*> snapmirror resync -destination-path svm3:vol1_dst_dst -source-path svm:vol1 Warning: All data newer than Snapshot copy snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_065900 on volume svm3:vol1_dst_dst will be deleted. Do you want to continue? {y|n}: y Operation is queued: initiate snapmirror resync to destination "svm3:vol1_dst_dst". ::*> snapmirror show Progress Source Destination Mirror Relationship Total Last Path Type Path State Status Progress Healthy Updated ----------- ---- ------------ ------- -------------- --------- ------- -------- svm:vol1 XDP svm2:vol1_dst Snapmirrored Idle - true - svm3:vol1_dst_dst Snapmirrored Idle - true - 2 entries were displayed. ::*> snapmirror show -destination-path svm3:vol1_dst_dst Source Path: svm:vol1 Source Cluster: - Source Vserver: svm Source Volume: vol1 Destination Path: svm3:vol1_dst_dst Destination Cluster: - Destination Vserver: svm3 Destination Volume: vol1_dst_dst Relationship Type: XDP Relationship Group Type: none Managing Vserver: svm3 SnapMirror Schedule: - SnapMirror Policy Type: async-mirror SnapMirror Policy: MirrorAllSnapshots Tries Limit: - Throttle (KB/sec): unlimited Consistency Group Item Mappings: - Current Transfer Throttle (KB/sec): - Mirror State: Snapmirrored Relationship Status: Idle File Restore File Count: - File Restore File List: - Transfer Snapshot: - Snapshot Progress: - Total Progress: - Network Compression Ratio: - Snapshot Checkpoint: - Newest Snapshot: snapmirror.674509fa-a065-11ee-981e-bdd56ead09c8_2148779290.2023-12-22_074801 Newest Snapshot Timestamp: 12/22 07:48:01 Exported Snapshot: snapmirror.674509fa-a065-11ee-981e-bdd56ead09c8_2148779290.2023-12-22_074801 Exported Snapshot Timestamp: 12/22 07:48:01 Healthy: true Relationship ID: 27f77b8a-a09e-11ee-981e-bdd56ead09c8 Source Vserver UUID: 04ee1778-a058-11ee-981e-bdd56ead09c8 Destination Vserver UUID: 674509fa-a065-11ee-981e-bdd56ead09c8 Current Operation ID: - Transfer Type: - Transfer Error: - Last Transfer Type: resync Last Transfer Error: - Last Transfer Error Codes: - Last Transfer Size: 3.27KB Last Transfer Network Compression Ratio: 1:1 Last Transfer Duration: 0:0:4 Last Transfer From: svm:vol1 Last Transfer End Timestamp: 12/22 07:48:05 Unhealthy Reason: - Progress Last Updated: - Relationship Capability: 8.2 and above Lag Time: 0:0:9 Current Transfer Priority: - SMTape Operation: - Destination Volume Node Name: FsxId0ab6f9b00824a187c-01 Identity Preserve Vserver DR: - Number of Successful Updates: 0 Number of Failed Updates: 1 Number of Successful Resyncs: 1 Number of Failed Resyncs: 0 Number of Successful Breaks: 0 Number of Failed Breaks: 0 Total Transfer Bytes: 3352 Total Transfer Time in Seconds: 4 Source Volume MSIDs Preserved: - OpMask: ffffffffffffffff Is Auto Expand Enabled: - Percent Complete for Current Status: -
再同期が完了しました。
Snapshotsnapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_065900
のデータは削除されたようです。
Storage Efficiency、ボリューム、aggregate、Snapshotの情報を確認します。
::*> volume efficiency show -volume vol1* -fields changelog-usage, changelog-size, logical-data-size, state, progress, last-op-size vserver volume state progress last-op-size changelog-usage changelog-size logical-data-size ------- ------ ------- ----------------- ------------ --------------- -------------- ----------------- svm vol1 Enabled Idle for 00:31:47 1.70GB 9% 15.41MB 8.11GB svm2 vol1_dst Enabled Idle for 00:19:39 12.21GB 0% 540KB 8.08GB svm3 vol1_dst_dst Disabled Idle for 00:16:20 83.62MB 0% 900KB 8.06GB 3 entries were displayed. ::*> volume efficiency inactive-data-compression show -volume vol1_dst_dst -instance Volume: vol1_dst_dst Vserver: svm3 Is Enabled: true Scan Mode: - Progress: IDLE Status: SUCCESS Compression Algorithm: lzopro Failure Reason: - Total Blocks: - Total blocks Processed: - Percentage: - Phase1 L1s Processed: - Phase1 Lns Skipped: - Phase2 Total Blocks: - Phase2 Blocks Processed: - Number of Cold Blocks Encountered: 0 Number of Repacked Blocks: 0 Number of Compression Done Blocks: 0 Number of Vol-Overwrites: 0 Time since Last Inactive Data Compression Scan started(sec): 952 Time since Last Inactive Data Compression Scan ended(sec): 951 Time since Last Successful Inactive Data Compression Scan started(sec): - Time since Last Successful Inactive Data Compression Scan ended(sec): 951 Average time for Cold Data Compression(sec): 0 Tuning Enabled: false Threshold: 0 Threshold Upper Limit: 0 Threshold Lower Limit: 0 Client Read history window: 0 Incompressible Data Percentage: 0% ::*> volume show -volume vol1* -fields available, filesystem-size, total, used, percent-used, performance-tier-inactive-user-data, performance-tier-inactive-user-data-percent, size, dedupe-space-saved, dedupe-space-saved-percent, dedupe-space-shared, logical-used, logical-used-percent,logical-used-by-afs, logical-available vserver volume size available filesystem-size total used percent-used dedupe-space-saved dedupe-space-saved-percent dedupe-space-shared logical-used logical-used-percent logical-available logical-used-by-afs performance-tier-inactive-user-data performance-tier-inactive-user-data-percent ------- ------ ---- --------- --------------- ------- ------ ------------ ------------------ -------------------------- ------------------- ------------ -------------------- ----------------- ------------------- ----------------------------------- ------------------------------------------- svm vol1 16GB 7.86GB 16GB 15.20GB 7.33GB 48% 791.6MB 10% 791.6MB 8.11GB 53% - 8.11GB - - svm2 vol1_dst 7.54GB 1.10GB 7.54GB 7.16GB 6.06GB 84% 4.97GB 45% 3.01GB 11.01GB 154% - 8.05GB 0B 0% svm3 vol1_dst_dst 5.24GB 844.8MB 5.24GB 5.24GB 4.41GB 84% 5.11GB 54% 2.21GB 9.52GB 182% - 8.06GB 0B 0% 3 entries were displayed. ::*> volume show-footprint -volume vol1_dst_dst Vserver : svm3 Volume : vol1_dst_dst Feature Used Used% -------------------------------- ---------- ----- Volume Data Footprint 4.41GB 0% Footprint in Performance Tier 4.51GB 100% Footprint in FSxFabricpoolObjectStore 0B 0% Volume Guarantee 0B 0% Flexible Volume Metadata 50.02MB 0% Delayed Frees 104.2MB 0% File Operation Metadata 4KB 0% Total Footprint 4.56GB 1% Effective Total Footprint 4.56GB 1% ::*> aggr show-efficiency -instance Name of the Aggregate: aggr1 Node where Aggregate Resides: FsxId0ab6f9b00824a187c-01 Logical Size Used by Volumes, Clones, Snapshot Copies in the Aggregate: 103.9GB Total Physical Used: 12.54GB Total Storage Efficiency Ratio: 8.29:1 Total Data Reduction Logical Used Without Snapshots: 24.07GB Total Data Reduction Physical Used Without Snapshots: 9.54GB Total Data Reduction Efficiency Ratio Without Snapshots: 2.52:1 Total Data Reduction Logical Used without snapshots and flexclones: 24.07GB Total Data Reduction Physical Used without snapshots and flexclones: 9.54GB Total Data Reduction Efficiency Ratio without snapshots and flexclones: 2.52:1 Total Logical Size Used by Volumes, Clones, Snapshot Copies in the FabricPool Performance Tier: 68.72GB Total Physical Used in FabricPool Performance Tier: 8.51GB Total FabricPool Performance Tier Storage Efficiency Ratio: 8.08:1 Total Data Reduction Logical Used without snapshots and flexclones in the FabricPool Performance Tier: 16.38GB Total Data Reduction Physical Used without snapshots and flexclones in the FabricPool Performance Tier: 5.53GB Total FabricPool Performance Tier Data Reduction Efficiency Ratio without snapshots and flexclones: 2.96:1 Logical Space Used for All Volumes: 24.07GB Physical Space Used for All Volumes: 13.21GB Space Saved by Volume Deduplication: 10.86GB Space Saved by Volume Deduplication and pattern detection: 10.86GB Volume Deduplication Savings ratio: 1.82:1 Space Saved by Volume Compression: 0B Volume Compression Savings ratio: 1.00:1 Space Saved by Inline Zero Pattern Detection: 0B Volume Data Reduction SE Ratio: 1.82:1 Logical Space Used by the Aggregate: 20.18GB Physical Space Used by the Aggregate: 12.54GB Space Saved by Aggregate Data Reduction: 7.64GB Aggregate Data Reduction SE Ratio: 1.61:1 Logical Size Used by Snapshot Copies: 79.86GB Physical Size Used by Snapshot Copies: 4.82GB Snapshot Volume Data Reduction Ratio: 16.57:1 Logical Size Used by FlexClone Volumes: 0B Physical Sized Used by FlexClone Volumes: 0B FlexClone Volume Data Reduction Ratio: 1.00:1 Snapshot And FlexClone Volume Data Reduction SE Ratio: 16.57:1 Number of Volumes Offline: 0 Number of SIS Disabled Volumes: 3 Number of SIS Change Log Disabled Volumes: 1 ::*> snapshot show -volume vol1* ---Blocks--- Vserver Volume Snapshot Size Total% Used% -------- -------- ------------------------------------- -------- ------ ----- svm vol1 test.2023-12-22_0533 160KB 0% 0% snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_053528 24.45MB 0% 0% snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_065900 272KB 0% 0% snapmirror.674509fa-a065-11ee-981e-bdd56ead09c8_2148779290.2023-12-22_074801 136KB 0% 0% svm2 vol1_dst snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_015623 1.91GB 25% 38% test.2023-12-22_0533 893.2MB 12% 22% snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_053528 481.9MB 6% 13% snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_065900 79.82MB 1% 2% svm3 vol1_dst_dst test.2023-12-22_0533 1.29GB 25% 30% snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_053528 101.9MB 2% 3% snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_065900 24.16MB 0% 1% snapmirror.674509fa-a065-11ee-981e-bdd56ead09c8_2148779290.2023-12-22_074801 41.10MB 1% 1% 12 entries were displayed.
重複排除量やaggregateレイヤーでのデータ削減量に大きな変化はありませんでした。再同期するにあたってせっかくデータ削減したものが削除されてしまうのかと思っていたのですが、そんなことはないようです。
Snapshotsnapmirror.674509fa-a065-11ee-981e-bdd56ead09c8_2148779290.2023-12-22_074801
のサイズは若干増加しています。
また、vol1_dst_dst
のStorage EfficiencyがDisableになっていました。snapmirror break
を実行してもStorage EfficinecyはEnableのままである認識です。そのため、snapmirror resync
のタイミングでDisableになったと考えます。
テスト用ファイルの追加
vol1
からvol1_dst_dst
で差分がある場合のSnapMirrorの動きが気になってきました。
vol1
にテスト用ファイルを追加します。
$ yes ABCDE | tr -d '\n' | sudo dd of=/mnt/fsxn/vol1/ABCDE_padding_file bs=1024 count=1024K 1048576+0 records in 1048576+0 records out 1073741824 bytes (1.1 GB, 1.0 GiB) copied, 7.32669 s, 147 MB/s $ ls -lh /mnt/fsxn/vol1/ABCDE_padding_file -rw-r--r--. 1 root root 1.0G Dec 22 07:53 /mnt/fsxn/vol1/ABCDE_padding_file $ df -hT -t nfs4 Filesystem Type Size Used Avail Use% Mounted on svm-0058ae83d258ab2e3.fs-0ab6f9b00824a187c.fsx.us-east-1.amazonaws.com:/vol1 nfs4 16G 8.4G 6.9G 55% /mnt/fsxn/vol1
svm と svm3 のボリューム間のSnapMirror update
svm と svm3 のボリューム間のSnapMirrorの差分転送を行います。
::*> snapmirror update -destination-path svm3:vol1_dst_dstOperation is queued: snapmirror update of destination "svm3:vol1_dst_dst". ::*> snapmirror show Progress Source Destination Mirror Relationship Total Last Path Type Path State Status Progress Healthy Updated ----------- ---- ------------ ------- -------------- --------- ------- -------- svm:vol1 XDP svm2:vol1_dst Snapmirrored Idle - true - svm3:vol1_dst_dst Snapmirrored Transferring 303.1MB true 12/22 07:55:40 2 entries were displayed. ::*> snapmirror show Progress Source Destination Mirror Relationship Total Last Path Type Path State Status Progress Healthy Updated ----------- ---- ------------ ------- -------------- --------- ------- -------- svm:vol1 XDP svm2:vol1_dst Snapmirrored Idle - true - svm3:vol1_dst_dst Snapmirrored Transferring 729.2MB true 12/22 07:56:27 2 entries were displayed. ::*> snapmirror show Progress Source Destination Mirror Relationship Total Last Path Type Path State Status Progress Healthy Updated ----------- ---- ------------ ------- -------------- --------- ------- -------- svm:vol1 XDP svm2:vol1_dst Snapmirrored Idle - true - svm3:vol1_dst_dst Snapmirrored Idle - true - 2 entries were displayed. ::*> snapmirror show -destination-path svm3:vol1_dst_dst Source Path: svm:vol1 Source Cluster: - Source Vserver: svm Source Volume: vol1 Destination Path: svm3:vol1_dst_dst Destination Cluster: - Destination Vserver: svm3 Destination Volume: vol1_dst_dst Relationship Type: XDP Relationship Group Type: none Managing Vserver: svm3 SnapMirror Schedule: - SnapMirror Policy Type: async-mirror SnapMirror Policy: MirrorAllSnapshots Tries Limit: - Throttle (KB/sec): unlimited Consistency Group Item Mappings: - Current Transfer Throttle (KB/sec): - Mirror State: Snapmirrored Relationship Status: Idle File Restore File Count: - File Restore File List: - Transfer Snapshot: - Snapshot Progress: - Total Progress: - Network Compression Ratio: - Snapshot Checkpoint: - Newest Snapshot: snapmirror.674509fa-a065-11ee-981e-bdd56ead09c8_2148779290.2023-12-22_075507 Newest Snapshot Timestamp: 12/22 07:55:07 Exported Snapshot: snapmirror.674509fa-a065-11ee-981e-bdd56ead09c8_2148779290.2023-12-22_075507 Exported Snapshot Timestamp: 12/22 07:55:07 Healthy: true Relationship ID: 27f77b8a-a09e-11ee-981e-bdd56ead09c8 Source Vserver UUID: 04ee1778-a058-11ee-981e-bdd56ead09c8 Destination Vserver UUID: 674509fa-a065-11ee-981e-bdd56ead09c8 Current Operation ID: - Transfer Type: - Transfer Error: - Last Transfer Type: update Last Transfer Error: - Last Transfer Error Codes: - Last Transfer Size: 1.02GB Last Transfer Network Compression Ratio: 1:1 Last Transfer Duration: 0:2:5 Last Transfer From: svm:vol1 Last Transfer End Timestamp: 12/22 07:57:12 Unhealthy Reason: - Progress Last Updated: - Relationship Capability: 8.2 and above Lag Time: 0:2:21 Current Transfer Priority: - SMTape Operation: - Destination Volume Node Name: FsxId0ab6f9b00824a187c-01 Identity Preserve Vserver DR: - Number of Successful Updates: 1 Number of Failed Updates: 1 Number of Successful Resyncs: 1 Number of Failed Resyncs: 0 Number of Successful Breaks: 0 Number of Failed Breaks: 0 Total Transfer Bytes: 1097840840 Total Transfer Time in Seconds: 129 Source Volume MSIDs Preserved: - OpMask: ffffffffffffffff Is Auto Expand Enabled: - Percent Complete for Current Status: -
転送量は1.02GBでした。きちんと差分のみを転送していることが分かります。
Storage Efficiency、ボリューム、aggregate、Snapshotの情報を確認します。
::*> volume efficiency show -volume vol1* -fields changelog-usage, changelog-size, logical-data-size, state, progress, last-op-size vserver volume state progress last-op-size changelog-usage changelog-size logical-data-size ------- ------ ------- ------------------ ------------ --------------- -------------- ----------------- svm vol1 Enabled 79120 KB (6%) Done 1.70GB 5% 25.41MB 9.12GB svm2 vol1_dst Enabled Idle for 00:28:45 12.21GB 0% 540KB 8.08GB svm3 vol1_dst_dst Disabled Idle for 00:25:26 83.62MB 0% 900KB 9.07GB 3 entries were displayed. ::*> volume efficiency inactive-data-compression show -volume vol1_dst_dst -instance Volume: vol1_dst_dst Vserver: svm3 Is Enabled: true Scan Mode: - Progress: IDLE Status: SUCCESS Compression Algorithm: lzopro Failure Reason: - Total Blocks: - Total blocks Processed: - Percentage: - Phase1 L1s Processed: - Phase1 Lns Skipped: - Phase2 Total Blocks: - Phase2 Blocks Processed: - Number of Cold Blocks Encountered: 0 Number of Repacked Blocks: 0 Number of Compression Done Blocks: 0 Number of Vol-Overwrites: 0 Time since Last Inactive Data Compression Scan started(sec): 1482 Time since Last Inactive Data Compression Scan ended(sec): 1481 Time since Last Successful Inactive Data Compression Scan started(sec): - Time since Last Successful Inactive Data Compression Scan ended(sec): 1481 Average time for Cold Data Compression(sec): 0 Tuning Enabled: false Threshold: 0 Threshold Upper Limit: 0 Threshold Lower Limit: 0 Client Read history window: 0 Incompressible Data Percentage: 0% ::*> volume show -volume vol1* -fields available, filesystem-size, total, used, percent-used, performance-tier-inactive-user-data, performance-tier-inactive-user-data-percent, size, dedupe-space-saved, dedupe-space-saved-percent, dedupe-space-shared, logical-used, logical-used-percent,logical-used-by-afs, logical-available vserver volume size available filesystem-size total used percent-used dedupe-space-saved dedupe-space-saved-percent dedupe-space-shared logical-used logical-used-percent logical-available logical-used-by-afs performance-tier-inactive-user-data performance-tier-inactive-user-data-percent ------- ------ ---- --------- --------------- ------- ------ ------------ ------------------ -------------------------- ------------------- ------------ -------------------- ----------------- ------------------- ----------------------------------- ------------------------------------------- svm vol1 16GB 6.85GB 16GB 15.20GB 8.35GB 54% 791.6MB 8% 791.6MB 9.12GB 60% - 9.12GB - - svm2 vol1_dst 7.54GB 1.10GB 7.54GB 7.16GB 6.06GB 84% 4.97GB 45% 3.01GB 11.01GB 154% - 8.05GB 0B 0% svm3 vol1_dst_dst 6.41GB 1010MB 6.41GB 6.41GB 5.42GB 84% 5.11GB 49% 3.21GB 10.53GB 164% - 9.07GB 0B 0% 3 entries were displayed. ::*> volume show-footprint -volume vol1_dst_dst Vserver : svm3 Volume : vol1_dst_dst Feature Used Used% -------------------------------- ---------- ----- Volume Data Footprint 5.42GB 1% Footprint in Performance Tier 5.50GB 100% Footprint in FSxFabricpoolObjectStore 0B 0% Volume Guarantee 0B 0% Flexible Volume Metadata 50.02MB 0% Delayed Frees 79.29MB 0% File Operation Metadata 4KB 0% Total Footprint 5.55GB 1% Effective Total Footprint 5.55GB 1% ::*> aggr show-efficiency -instance Name of the Aggregate: aggr1 Node where Aggregate Resides: FsxId0ab6f9b00824a187c-01 Logical Size Used by Volumes, Clones, Snapshot Copies in the Aggregate: 116.0GB Total Physical Used: 11.69GB Total Storage Efficiency Ratio: 9.92:1 Total Data Reduction Logical Used Without Snapshots: 26.08GB Total Data Reduction Physical Used Without Snapshots: 8.74GB Total Data Reduction Efficiency Ratio Without Snapshots: 2.98:1 Total Data Reduction Logical Used without snapshots and flexclones: 26.08GB Total Data Reduction Physical Used without snapshots and flexclones: 8.74GB Total Data Reduction Efficiency Ratio without snapshots and flexclones: 2.98:1 Total Logical Size Used by Volumes, Clones, Snapshot Copies in the FabricPool Performance Tier: 78.82GB Total Physical Used in FabricPool Performance Tier: 7.67GB Total FabricPool Performance Tier Storage Efficiency Ratio: 10.28:1 Total Data Reduction Logical Used without snapshots and flexclones in the FabricPool Performance Tier: 17.41GB Total Data Reduction Physical Used without snapshots and flexclones in the FabricPool Performance Tier: 4.73GB Total FabricPool Performance Tier Data Reduction Efficiency Ratio without snapshots and flexclones: 3.68:1 Logical Space Used for All Volumes: 26.08GB Physical Space Used for All Volumes: 15.22GB Space Saved by Volume Deduplication: 10.86GB Space Saved by Volume Deduplication and pattern detection: 10.86GB Volume Deduplication Savings ratio: 1.71:1 Space Saved by Volume Compression: 0B Volume Compression Savings ratio: 1.00:1 Space Saved by Inline Zero Pattern Detection: 0B Volume Data Reduction SE Ratio: 1.71:1 Logical Space Used by the Aggregate: 19.08GB Physical Space Used by the Aggregate: 11.69GB Space Saved by Aggregate Data Reduction: 7.39GB Aggregate Data Reduction SE Ratio: 1.63:1 Logical Size Used by Snapshot Copies: 89.94GB Physical Size Used by Snapshot Copies: 4.82GB Snapshot Volume Data Reduction Ratio: 18.66:1 Logical Size Used by FlexClone Volumes: 0B Physical Sized Used by FlexClone Volumes: 0B FlexClone Volume Data Reduction Ratio: 1.00:1 Snapshot And FlexClone Volume Data Reduction SE Ratio: 18.66:1 Number of Volumes Offline: 0 Number of SIS Disabled Volumes: 3 Number of SIS Change Log Disabled Volumes: 1 ::*> snapshot show -volume vol1* ---Blocks--- Vserver Volume Snapshot Size Total% Used% -------- -------- ------------------------------------- -------- ------ ----- svm vol1 test.2023-12-22_0533 160KB 0% 0% snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_053528 24.45MB 0% 0% snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_065900 312KB 0% 0% snapmirror.674509fa-a065-11ee-981e-bdd56ead09c8_2148779290.2023-12-22_075507 168KB 0% 0% svm2 vol1_dst snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_015623 1.91GB 25% 38% test.2023-12-22_0533 893.2MB 12% 22% snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_053528 481.9MB 6% 13% snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_065900 79.82MB 1% 2% svm3 vol1_dst_dst test.2023-12-22_0533 1.29GB 20% 25% snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_053528 101.9MB 2% 2% snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_065900 24.16MB 0% 1% snapmirror.674509fa-a065-11ee-981e-bdd56ead09c8_2148779290.2023-12-22_074801 41.57MB 1% 1% snapmirror.674509fa-a065-11ee-981e-bdd56ead09c8_2148779290.2023-12-22_075507 160KB 0% 0% 13 entries were displayed. ::*> volume efficiency show -volume vol1* -fields state, policy, storage-efficiency-mode, inline-compression, inline-dedupe, compression, data-compaction, auto-adaptive-compression-savings, auto-adaptive-compression-existing-volume,using-auto-adaptive-compression vserver volume state policy compression inline-compression storage-efficiency-mode inline-dedupe data-compaction auto-adaptive-compression-savings using-auto-adaptive-compression auto-adaptive-compression-existing-volume ------- ------ ------- ------ ----------- ------------------ ----------------------- ------------- --------------- --------------------------------- ------------------------------- ----------------------------------------- svm vol1 Enabled auto false true efficient false true true true false svm2 vol1_dst Enabled - false true efficient false true true true false svm3 vol1_dst_dst Disabled - false true efficient false true true true false 3 entries were displayed.
重複排除量やaggregateレイヤーでのデータ削減量に大きな変化はありませんでした。
また、DisableだからかSnapMirror転送後にvol1_dst_dst
でのStorage Efficinecyは実行されていないようでした。SnapMirror initializeやresync後にStorage Efficiencyを有効化した方が良いかもしれません。
vol1_dst_dst のSnapshotを削除する
svm2 と svm3 のボリューム間のSnapMirrorのカットオーバー
vol1_dst_dst
のSnapshotを削除して、ボリュームやaggregateの使用量が減ることを確認します。
下準備として、vm2 と svm3 のボリューム間のSnapMirrorをカットオーバーします。
::*> snapmirror quiesce -destination-path svm3:vol1_dst_dst Operation succeeded: snapmirror quiesce for destination "svm3:vol1_dst_dst". ::*> snapmirror break -destination-path svm3:vol1_dst_dst Operation succeeded: snapmirror break for destination "svm3:vol1_dst_dst". ::*> snapmirror show Progress Source Destination Mirror Relationship Total Last Path Type Path State Status Progress Healthy Updated ----------- ---- ------------ ------- -------------- --------- ------- -------- svm:vol1 XDP svm2:vol1_dst Snapmirrored Idle - true - svm3:vol1_dst_dst Broken-off Idle - true - 2 entries were displayed.
Storage Efficiencyの情報を確認します。
::*> volume efficiency show -volume vol1* -fields state, policy, storage-efficiency-mode, inline-compression, inline-dedupe, compression, data-compaction, auto-adaptive-compression-savings, auto-adaptive-compression-existing-volume,using-auto-adaptive-compression vserver volume state policy compression inline-compression storage-efficiency-mode inline-dedupe data-compaction auto-adaptive-compression-savings using-auto-adaptive-compression auto-adaptive-compression-existing-volume ------- ------ ------- ------ ----------- ------------------ ----------------------- ------------- --------------- --------------------------------- ------------------------------- ----------------------------------------- svm vol1 Enabled auto false true efficient false true true true false svm2 vol1_dst Enabled - false true efficient false true true true false svm3 vol1_dst_dst Enabled auto false true efficient true true true true false 3 entries were displayed. ::*> volume efficiency inactive-data-compression show -volume vol1_dst_dst -instance Volume: vol1_dst_dst Vserver: svm3 Is Enabled: true Scan Mode: - Progress: IDLE Status: SUCCESS Compression Algorithm: lzopro Failure Reason: - Total Blocks: - Total blocks Processed: - Percentage: - Phase1 L1s Processed: - Phase1 Lns Skipped: - Phase2 Total Blocks: - Phase2 Blocks Processed: - Number of Cold Blocks Encountered: 0 Number of Repacked Blocks: 0 Number of Compression Done Blocks: 0 Number of Vol-Overwrites: 0 Time since Last Inactive Data Compression Scan started(sec): 1920 Time since Last Inactive Data Compression Scan ended(sec): 1919 Time since Last Successful Inactive Data Compression Scan started(sec): - Time since Last Successful Inactive Data Compression Scan ended(sec): 1919 Average time for Cold Data Compression(sec): 0 Tuning Enabled: true Threshold: 14 Threshold Upper Limit: 21 Threshold Lower Limit: 14 Client Read history window: 14 Incompressible Data Percentage: 0%
Storage Efficiencyが有効になりました。また、ポリシーがauto
になったり、インライン重複排除が有効になったりしています。
Snapshotの削除
vol1_dst_dst
のSnapshotを少しづつ削除していきます。
Snapshotの一覧を確認します。
::*> snapshot show -volume vol1* ---Blocks--- Vserver Volume Snapshot Size Total% Used% -------- -------- ------------------------------------- -------- ------ ----- svm vol1 test.2023-12-22_0533 160KB 0% 0% snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_053528 24.45MB 0% 0% snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_065900 312KB 0% 0% snapmirror.674509fa-a065-11ee-981e-bdd56ead09c8_2148779290.2023-12-22_075507 168KB 0% 0% svm2 vol1_dst snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_015623 1.91GB 25% 38% test.2023-12-22_0533 893.2MB 12% 22% snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_053528 481.9MB 6% 13% snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_065900 79.82MB 1% 2% svm3 vol1_dst_dst test.2023-12-22_0533 1.29GB 20% 25% snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_053528 101.9MB 2% 2% snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_065900 24.16MB 0% 1% snapmirror.674509fa-a065-11ee-981e-bdd56ead09c8_2148779290.2023-12-22_074801 41.57MB 1% 1% snapmirror.674509fa-a065-11ee-981e-bdd56ead09c8_2148779290.2023-12-22_075507 4.94MB 0% 0% 13 entries were displayed.
最も古いtest.2023-12-22_0533
を削除します。
::*> snapshot delete -vserver svm3 -volume vol1_dst_dst -snapshot test.2023-12-22_0533 Warning: Deleting a Snapshot copy permanently removes data that is stored only in that Snapshot copy. Are you sure you want to delete Snapshot copy "test.2023-12-22_0533" for volume "vol1_dst_dst" in Vserver "svm3" ? {y|n}: y ::*> snapshot show -volume vol1* ---Blocks--- Vserver Volume Snapshot Size Total% Used% -------- -------- ------------------------------------- -------- ------ ----- svm vol1 test.2023-12-22_0533 160KB 0% 0% snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_053528 24.45MB 0% 0% snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_065900 312KB 0% 0% snapmirror.674509fa-a065-11ee-981e-bdd56ead09c8_2148779290.2023-12-22_075507 168KB 0% 0% svm2 vol1_dst snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_015623 1.91GB 25% 38% test.2023-12-22_0533 893.2MB 12% 22% snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_053528 481.9MB 6% 13% snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_065900 79.82MB 1% 2% svm3 vol1_dst_dst snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_053528 101.9MB 2% 2% snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_065900 24.16MB 0% 1% snapmirror.674509fa-a065-11ee-981e-bdd56ead09c8_2148779290.2023-12-22_074801 41.57MB 1% 1% snapmirror.674509fa-a065-11ee-981e-bdd56ead09c8_2148779290.2023-12-22_075507 4.95MB 0% 0% 12 entries were displayed.
ボリューム、aggregateの情報を確認します。
::*> volume show -volume vol1* -fields available, filesystem-size, total, used, percent-used, performance-tier-inactive-user-data, performance-tier-inactive-user-data-percent, size, dedupe-space-saved, dedupe-space-saved-percent, dedupe-space-shared, logical-used, logical-used-percent,logical-used-by-afs, logical-available vserver volume size available filesystem-size total used percent-used dedupe-space-saved dedupe-space-saved-percent dedupe-space-shared logical-used logical-used-percent logical-available logical-used-by-afs performance-tier-inactive-user-data performance-tier-inactive-user-data-percent ------- ------ ---- --------- --------------- ------- ------ ------------ ------------------ -------------------------- ------------------- ------------ -------------------- ----------------- ------------------- ----------------------------------- ------------------------------------------- svm vol1 16GB 6.85GB 16GB 15.20GB 8.35GB 54% 791.6MB 8% 791.6MB 9.12GB 60% - 9.12GB - - svm2 vol1_dst 7.54GB 1.10GB 7.54GB 7.16GB 6.06GB 84% 4.97GB 45% 3.01GB 11.01GB 154% - 8.05GB 0B 0% svm3 vol1_dst_dst 4.90GB 795.2MB 4.90GB 4.90GB 4.12GB 84% 5.11GB 55% 2.21GB 9.24GB 188% - 9.07GB 0B 0% 3 entries were displayed. ::*> volume show-footprint -volume vol1_dst_dst Vserver : svm3 Volume : vol1_dst_dst Feature Used Used% -------------------------------- ---------- ----- Volume Data Footprint 4.12GB 0% Footprint in Performance Tier 5.07GB 100% Footprint in FSxFabricpoolObjectStore 0B 0% Volume Guarantee 0B 0% Flexible Volume Metadata 0B 0% Delayed Frees 969.2MB 0% File Operation Metadata 4KB 0% Total Footprint 5.07GB 1% Effective Total Footprint 5.07GB 1% ::*> aggr show-efficiency -instance Name of the Aggregate: aggr1 Node where Aggregate Resides: FsxId0ab6f9b00824a187c-01 Logical Size Used by Volumes, Clones, Snapshot Copies in the Aggregate: 109.9GB Total Physical Used: 11.83GB Total Storage Efficiency Ratio: 9.29:1 Total Data Reduction Logical Used Without Snapshots: 26.08GB Total Data Reduction Physical Used Without Snapshots: 9.58GB Total Data Reduction Efficiency Ratio Without Snapshots: 2.72:1 Total Data Reduction Logical Used without snapshots and flexclones: 26.08GB Total Data Reduction Physical Used without snapshots and flexclones: 9.58GB Total Data Reduction Efficiency Ratio without snapshots and flexclones: 2.72:1 Total Logical Size Used by Volumes, Clones, Snapshot Copies in the FabricPool Performance Tier: 72.70GB Total Physical Used in FabricPool Performance Tier: 7.80GB Total FabricPool Performance Tier Storage Efficiency Ratio: 9.32:1 Total Data Reduction Logical Used without snapshots and flexclones in the FabricPool Performance Tier: 17.40GB Total Data Reduction Physical Used without snapshots and flexclones in the FabricPool Performance Tier: 5.57GB Total FabricPool Performance Tier Data Reduction Efficiency Ratio without snapshots and flexclones: 3.13:1 Logical Space Used for All Volumes: 26.08GB Physical Space Used for All Volumes: 15.22GB Space Saved by Volume Deduplication: 10.86GB Space Saved by Volume Deduplication and pattern detection: 10.86GB Volume Deduplication Savings ratio: 1.71:1 Space Saved by Volume Compression: 0B Volume Compression Savings ratio: 1.00:1 Space Saved by Inline Zero Pattern Detection: 0B Volume Data Reduction SE Ratio: 1.71:1 Logical Space Used by the Aggregate: 18.56GB Physical Space Used by the Aggregate: 11.83GB Space Saved by Aggregate Data Reduction: 6.73GB Aggregate Data Reduction SE Ratio: 1.57:1 Logical Size Used by Snapshot Copies: 83.82GB Physical Size Used by Snapshot Copies: 3.53GB Snapshot Volume Data Reduction Ratio: 23.76:1 Logical Size Used by FlexClone Volumes: 0B Physical Sized Used by FlexClone Volumes: 0B FlexClone Volume Data Reduction Ratio: 1.00:1 Snapshot And FlexClone Volume Data Reduction SE Ratio: 23.76:1 Number of Volumes Offline: 0 Number of SIS Disabled Volumes: 3 Number of SIS Change Log Disabled Volumes: 0
logical-used
が10.53GBから9.24GBと減っています。
また、Physical Size Used by Snapshot Copies
が4.82GBから3.53GBとSnapshotの物理的な使用量が削減されています。
NFSクライアントからvol1_dst_dst
の使用量を確認
vol1_dst_dst
をNFSでマウントして、どのようなサイズでレポートされるのか確認します。
まず、vol1_dst_dst
を/vol1_dst_dst
にマウントします。
::*> volume mount -vserver svm3 -volume vol1_dst_dst -junction-path /vol1_dst_dst Queued private job: 32
NFSクライアントからこちらのボリュームをマウントします。
$ sudo mount -t nfs svm-00c880c6c6eb922ed.fs-0ab6f9b00824a187c.fsx.us-east-1.amazonaws.com:/vol1_dst_dst /mnt/fsxn/vol1_dst_dst/ $ ls -lh /mnt/fsxn/vol1_dst_dst/ total 9.1G -rw-r--r--. 1 root root 1.0G Dec 22 01:46 1_padding_file -rw-r--r--. 1 root root 1.0G Dec 22 07:53 ABCDE_padding_file -rw-r--r--. 1 root root 1.0G Dec 22 05:28 a_padding_file -rw-r--r--. 1 root root 1.0G Dec 22 06:55 abcde_padding_file -rw-r--r--. 1 root root 1.0G Dec 22 01:47 urandom_block_file -rw-r--r--. 1 root root 1.0G Dec 22 05:02 urandom_block_file2 -rw-r--r--. 1 root root 1.0G Dec 22 05:02 urandom_block_file2_copy -rw-r--r--. 1 root root 1.0G Dec 22 01:47 urandom_block_file_copy -rw-r--r--. 1 root root 1.0G Dec 22 06:41 urandom_block_file_copy2 $ df -hT -t nfs4 Filesystem Type Size Used Avail Use% Mounted on svm-0058ae83d258ab2e3.fs-0ab6f9b00824a187c.fsx.us-east-1.amazonaws.com:/vol1 nfs4 16G 8.4G 6.9G 55% /mnt/fsxn/vol1 svm-00c880c6c6eb922ed.fs-0ab6f9b00824a187c.fsx.us-east-1.amazonaws.com:/vol1_dst_dst nfs4 5.0G 4.2G 796M 85% /mnt/fsxn/vol1_dst_dst
使用量は4.2Gでレポートされています。物理的な使用量としてレポートされていることが分かります。
Snapshotをさらに削除
さらにSnasphotを削除します。対象はサイズが101MBのSnapshotsnapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_053528
です。
::*> snapshot delete -vserver svm3 -volume vol1_dst_dst -snapshot snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_053528 Warning: Deleting a Snapshot copy permanently removes data that is stored only in that Snapshot copy. Are you sure you want to delete Snapshot copy "snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_053528" for volume "vol1_dst_dst" in Vserver "svm3" ? {y|n}: y ::*> snapshot show -volume vol1* ---Blocks--- Vserver Volume Snapshot Size Total% Used% -------- -------- ------------------------------------- -------- ------ ----- svm vol1 test.2023-12-22_0533 160KB 0% 0% snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_053528 24.45MB 0% 0% snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_065900 312KB 0% 0% snapmirror.674509fa-a065-11ee-981e-bdd56ead09c8_2148779290.2023-12-22_075507 288KB 0% 0% svm2 vol1_dst snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_015623 1.91GB 25% 38% test.2023-12-22_0533 893.2MB 12% 22% snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_053528 481.9MB 6% 13% snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_065900 79.82MB 1% 2% svm3 vol1_dst_dst snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_065900 24.16MB 0% 1% snapmirror.674509fa-a065-11ee-981e-bdd56ead09c8_2148779290.2023-12-22_074801 41.57MB 1% 1% snapmirror.674509fa-a065-11ee-981e-bdd56ead09c8_2148779290.2023-12-22_075507 4.96MB 0% 0% 11 entries were displayed. ::*> volume show -volume vol1* -fields available, filesystem-size, total, used, percent-used, performance-tier-inactive-user-data, performance-tier-inactive-user-data-percent, size, dedupe-space-saved, dedupe-space-saved-percent, dedupe-space-shared, logical-used, logical-used-percent,logical-used-by-afs, logical-available vserver volume size available filesystem-size total used percent-used dedupe-space-saved dedupe-space-saved-percent dedupe-space-shared logical-used logical-used-percent logical-available logical-used-by-afs performance-tier-inactive-user-data performance-tier-inactive-user-data-percent ------- ------ ---- --------- --------------- ------- ------ ------------ ------------------ -------------------------- ------------------- ------------ -------------------- ----------------- ------------------- ----------------------------------- ------------------------------------------- svm vol1 16GB 6.85GB 16GB 15.20GB 8.35GB 54% 808.8MB 9% 808.8MB 9.14GB 60% - 9.14GB - - svm2 vol1_dst 7.54GB 1.10GB 7.54GB 7.16GB 6.06GB 84% 4.97GB 45% 3.01GB 11.01GB 154% - 8.05GB 0B 0% svm3 vol1_dst_dst 4.90GB 896.8MB 4.90GB 4.90GB 4.03GB 82% 5.11GB 56% 2.21GB 9.14GB 186% - 9.07GB 0B 0% 3 entries were displayed. ::*> volume show-footprint -volume vol1_dst_dst Vserver : svm3 Volume : vol1_dst_dst Feature Used Used% -------------------------------- ---------- ----- Volume Data Footprint 4.03GB 0% Footprint in Performance Tier 4.17GB 100% Footprint in FSxFabricpoolObjectStore 0B 0% Volume Guarantee 0B 0% Flexible Volume Metadata 50.02MB 0% Delayed Frees 145.3MB 0% File Operation Metadata 4KB 0% Total Footprint 4.22GB 0% Effective Total Footprint 4.22GB 0% ::*> aggr show-efficiency -instance Name of the Aggregate: aggr1 Node where Aggregate Resides: FsxId0ab6f9b00824a187c-01 Logical Size Used by Volumes, Clones, Snapshot Copies in the Aggregate: 103.9GB Total Physical Used: 11.04GB Total Storage Efficiency Ratio: 9.41:1 Total Data Reduction Logical Used Without Snapshots: 26.09GB Total Data Reduction Physical Used Without Snapshots: 8.78GB Total Data Reduction Efficiency Ratio Without Snapshots: 2.97:1 Total Data Reduction Logical Used without snapshots and flexclones: 26.09GB Total Data Reduction Physical Used without snapshots and flexclones: 8.78GB Total Data Reduction Efficiency Ratio without snapshots and flexclones: 2.97:1 Total Logical Size Used by Volumes, Clones, Snapshot Copies in the FabricPool Performance Tier: 66.66GB Total Physical Used in FabricPool Performance Tier: 7.01GB Total FabricPool Performance Tier Storage Efficiency Ratio: 9.51:1 Total Data Reduction Logical Used without snapshots and flexclones in the FabricPool Performance Tier: 17.41GB Total Data Reduction Physical Used without snapshots and flexclones in the FabricPool Performance Tier: 4.77GB Total FabricPool Performance Tier Data Reduction Efficiency Ratio without snapshots and flexclones: 3.65:1 Logical Space Used for All Volumes: 26.09GB Physical Space Used for All Volumes: 15.22GB Space Saved by Volume Deduplication: 10.87GB Space Saved by Volume Deduplication and pattern detection: 10.87GB Volume Deduplication Savings ratio: 1.71:1 Space Saved by Volume Compression: 0B Volume Compression Savings ratio: 1.00:1 Space Saved by Inline Zero Pattern Detection: 0B Volume Data Reduction SE Ratio: 1.71:1 Logical Space Used by the Aggregate: 16.74GB Physical Space Used by the Aggregate: 11.04GB Space Saved by Aggregate Data Reduction: 5.70GB Aggregate Data Reduction SE Ratio: 1.52:1 Logical Size Used by Snapshot Copies: 77.78GB Physical Size Used by Snapshot Copies: 3.43GB Snapshot Volume Data Reduction Ratio: 22.68:1 Logical Size Used by FlexClone Volumes: 0B Physical Sized Used by FlexClone Volumes: 0B FlexClone Volume Data Reduction Ratio: 1.00:1 Snapshot And FlexClone Volume Data Reduction SE Ratio: 22.68:1 Number of Volumes Offline: 0 Number of SIS Disabled Volumes: 3 Number of SIS Change Log Disabled Volumes: 0
logical-used
が9.24GBから9.14GBと減っています。
NFSクライアントからも使用量を確認します。
$ df -hT -t nfs4 Filesystem Type Size Used Avail Use% Mounted on svm-0058ae83d258ab2e3.fs-0ab6f9b00824a187c.fsx.us-east-1.amazonaws.com:/vol1 nfs4 16G 8.4G 6.9G 55% /mnt/fsxn/vol1 svm-00c880c6c6eb922ed.fs-0ab6f9b00824a187c.fsx.us-east-1.amazonaws.com:/vol1_dst_dst nfs4 5.0G 4.1G 897M 83% /mnt/fsxn/vol1_dst_dst
100MBほど減っていますね。
3日弱何もせずに放置してみます。この状態でボリューム、aggregate、Snapshotの情報を確認します。
b6f9b00824a187::*> volume show -volume vol1* -fields available, filesystem-size, total, used, percent-used, performance-tier-inactive-user-data, performance-tier-inactive-user-data-percent, size, dedupe-space-saved, dedupe-space-saved-percent, dedupe-space-shared,logical-used, logical-used-percent,logical-used-by-afs, logical-available vserver volume size available filesystem-size total used percent-used dedupe-space-saved dedupe-space-saved-percent dedupe-space-shared logical-used logical-used-percent logical-available logical-used-by-afs performance-tier-inactive-user-data performance-tier-inactive-user-data-percent ------- ------ ---- --------- --------------- ------- ------ ------------ ------------------ -------------------------- ------------------- ------------ -------------------- ----------------- ------------------- ----------------------------------- ------------------------------------------- svm vol1 16GB 6.85GB 16GB 15.20GB 8.35GB 54% 808.8MB 9% 808.8MB 9.14GB 60% - 9.14GB - - svm2 vol1_dst 7.54GB 1.10GB 7.54GB 7.16GB 6.06GB 84% 4.97GB 45% 3.01GB 11.01GB 154% - 8.05GB 0B 0% svm3 vol1_dst_dst 4.90GB 896.4MB 4.90GB 4.90GB 4.03GB 82% 5.11GB 56% 2.21GB 9.14GB 186% - 9.07GB 0B 0% 3 entries were displayed. ::*> volume show-footprint -volume vol1_dst_dst Vserver : svm3 Volume : vol1_dst_dst Feature Used Used% -------------------------------- ---------- ----- Volume Data Footprint 4.03GB 0% Footprint in Performance Tier 4.03GB 100% Footprint in FSxFabricpoolObjectStore 0B 0% Volume Guarantee 0B 0% Flexible Volume Metadata 50.02MB 0% Delayed Frees 1.34MB 0% File Operation Metadata 4KB 0% Total Footprint 4.08GB 0% Footprint Data Reduction 641.0MB 0% Auto Adaptive Compression 641.0MB 0% Effective Total Footprint 3.45GB 0% ::*> aggr show-efficiency -instance Name of the Aggregate: aggr1 Node where Aggregate Resides: FsxId0ab6f9b00824a187c-01 Logical Size Used by Volumes, Clones, Snapshot Copies in the Aggregate: 103.8GB Total Physical Used: 11.52GB Total Storage Efficiency Ratio: 9.02:1 Total Data Reduction Logical Used Without Snapshots: 25.94GB Total Data Reduction Physical Used Without Snapshots: 9.19GB Total Data Reduction Efficiency Ratio Without Snapshots: 2.82:1 Total Data Reduction Logical Used without snapshots and flexclones: 25.94GB Total Data Reduction Physical Used without snapshots and flexclones: 9.19GB Total Data Reduction Efficiency Ratio without snapshots and flexclones: 2.82:1 Total Logical Size Used by Volumes, Clones, Snapshot Copies in the FabricPool Performance Tier: 66.80GB Total Physical Used in FabricPool Performance Tier: 7.67GB Total FabricPool Performance Tier Storage Efficiency Ratio: 8.71:1 Total Data Reduction Logical Used without snapshots and flexclones in the FabricPool Performance Tier: 17.42GB Total Data Reduction Physical Used without snapshots and flexclones in the FabricPool Performance Tier: 5.35GB Total FabricPool Performance Tier Data Reduction Efficiency Ratio without snapshots and flexclones: 3.25:1 Logical Space Used for All Volumes: 25.94GB Physical Space Used for All Volumes: 15.07GB Space Saved by Volume Deduplication: 10.87GB Space Saved by Volume Deduplication and pattern detection: 10.87GB Volume Deduplication Savings ratio: 1.72:1 Space Saved by Volume Compression: 0B Volume Compression Savings ratio: 1.00:1 Space Saved by Inline Zero Pattern Detection: 0B Volume Data Reduction SE Ratio: 1.72:1 Logical Space Used by the Aggregate: 16.97GB Physical Space Used by the Aggregate: 11.52GB Space Saved by Aggregate Data Reduction: 5.45GB Aggregate Data Reduction SE Ratio: 1.47:1 Logical Size Used by Snapshot Copies: 77.89GB Physical Size Used by Snapshot Copies: 3.43GB Snapshot Volume Data Reduction Ratio: 22.71:1 Logical Size Used by FlexClone Volumes: 0B Physical Sized Used by FlexClone Volumes: 0B FlexClone Volume Data Reduction Ratio: 1.00:1 Snapshot And FlexClone Volume Data Reduction SE Ratio: 22.71:1 Number of Volumes Offline: 0 Number of SIS Disabled Volumes: 3 Number of SIS Change Log Disabled Volumes: 0 ::*> snapshot show -volume vol1* ---Blocks--- Vserver Volume Snapshot Size Total% Used% -------- -------- ------------------------------------- -------- ------ ----- svm vol1 test.2023-12-22_0533 160KB 0% 0% snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_053528 24.45MB 0% 0% snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_065900 312KB 0% 0% snapmirror.674509fa-a065-11ee-981e-bdd56ead09c8_2148779290.2023-12-22_075507 292KB 0% 0% svm2 vol1_dst snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_015623 1.91GB 25% 38% test.2023-12-22_0533 893.2MB 12% 22% snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_053528 481.9MB 6% 13% snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_065900 79.82MB 1% 2% svm3 vol1_dst_dst snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_065900 24.16MB 0% 1% snapmirror.674509fa-a065-11ee-981e-bdd56ead09c8_2148779290.2023-12-22_074801 41.57MB 1% 1% snapmirror.674509fa-a065-11ee-981e-bdd56ead09c8_2148779290.2023-12-22_075507 4.96MB 0% 0% 11 entries were displayed. ::*> volume efficiency inactive-data-compression show -volume vol1_dst_dst -instance Volume: vol1_dst_dst Vserver: svm3 Is Enabled: true Scan Mode: - Progress: IDLE Status: SUCCESS Compression Algorithm: lzopro Failure Reason: - Total Blocks: - Total blocks Processed: - Percentage: - Phase1 L1s Processed: - Phase1 Lns Skipped: - Phase2 Total Blocks: - Phase2 Blocks Processed: - Number of Cold Blocks Encountered: 0 Number of Repacked Blocks: 0 Number of Compression Done Blocks: 0 Number of Vol-Overwrites: 760 Time since Last Inactive Data Compression Scan started(sec): 0 Time since Last Inactive Data Compression Scan ended(sec): 12881 Time since Last Successful Inactive Data Compression Scan started(sec): - Time since Last Successful Inactive Data Compression Scan ended(sec): 0 Average time for Cold Data Compression(sec): 0 Tuning Enabled: true Threshold: 14 Threshold Upper Limit: 21 Threshold Lower Limit: 14 Client Read history window: 14 Incompressible Data Percentage: 0%
volume show-footprint
でAuto Adaptive Compression
が641.0MBと表示されていますね。ただし、Space Saved by Aggregate Data Reduction
は5.70GBから5.45GBとむしろ減っていました。
カスケードSnapMirrorでStorage Efficinecyを維持できる条件を整理する
svm と svm2 のボリューム間のSnapMirror initialize
カスケードSnapMirrorでStorage Efficinecyを維持できる条件を整理したいと思います。
先ほどの検証ではvol1 -> vol1_dst -> vol1_dst_dst
とカスケードSnapMirrorを2回繰り返すと、vol1_dst
で実行したStorage Efficiencyの効果をvol1_dst_dst
で維持することができました。
ここでvol1_dst -> vol1_dst_dst
のみ転送した場合の挙動が気になりました。
まず、svm と svm2 のボリューム間のSnapMirror relationshipのinitializeを行います。
転送元のボリュームはvol1
で、転送先のボリュームはvol1_dst2
と新しいものを用意します。
::*> snapmirror protect -path-list svm:vol1 -destination-vserver svm2 -policy MirrorAllSnapshots -auto-initialize true -support-tiering true -tiering-policy none -destination-volume-suffix _dst2 [Job 106] Job is queued: snapmirror protect for list of source endpoints beginning with "svm:vol1". ::*> volume show -vserver svm2 Vserver Volume Aggregate State Type Size Available Used% --------- ------------ ------------ ---------- ---- ---------- ---------- ----- svm2 svm2_root aggr1 online RW 1GB 967.3MB 0% svm2 vol1_dst aggr1 online DP 7.54GB 1.10GB 84% svm2 vol1_dst2 aggr1 online DP 9.21GB 7.88GB 9% 3 entries were displayed. ::*> snapmirror show Progress Source Destination Mirror Relationship Total Last Path Type Path State Status Progress Healthy Updated ----------- ---- ------------ ------- -------------- --------- ------- -------- svm:vol1 XDP svm2:vol1_dst Snapmirrored Idle - true - svm2:vol1_dst2 Uninitialized Transferring 1.08GB true 12/25 05:04:32 svm3:vol1_dst_dst Broken-off Idle - true - 3 entries were displayed. ::*> snapmirror show Progress Source Destination Mirror Relationship Total Last Path Type Path State Status Progress Healthy Updated ----------- ---- ------------ ------- -------------- --------- ------- -------- svm:vol1 XDP svm2:vol1_dst Snapmirrored Idle - true - svm2:vol1_dst2 Snapmirrored Finalizing 3.07GB true 12/25 05:07:45 svm3:vol1_dst_dst Broken-off Idle - true - 3 entries were displayed. ::*> snapmirror show Progress Source Destination Mirror Relationship Total Last Path Type Path State Status Progress Healthy Updated ----------- ---- ------------ ------- -------------- --------- ------- -------- svm:vol1 XDP svm2:vol1_dst Snapmirrored Idle - true - svm2:vol1_dst2 Snapmirrored Idle - true - svm3:vol1_dst_dst Broken-off Idle - true - 3 entries were displayed. ::*> snapmirror show -destination-path svm2:vol1_dst2 Source Path: svm:vol1 Source Cluster: - Source Vserver: svm Source Volume: vol1 Destination Path: svm2:vol1_dst2 Destination Cluster: - Destination Vserver: svm2 Destination Volume: vol1_dst2 Relationship Type: XDP Relationship Group Type: none Managing Vserver: svm2 SnapMirror Schedule: - SnapMirror Policy Type: async-mirror SnapMirror Policy: MirrorAllSnapshots Tries Limit: - Throttle (KB/sec): unlimited Consistency Group Item Mappings: - Current Transfer Throttle (KB/sec): - Mirror State: Snapmirrored Relationship Status: Idle File Restore File Count: - File Restore File List: - Transfer Snapshot: - Snapshot Progress: - Total Progress: - Network Compression Ratio: - Snapshot Checkpoint: - Newest Snapshot: snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779291.2023-12-25_050406 Newest Snapshot Timestamp: 12/25 05:04:06 Exported Snapshot: snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779291.2023-12-25_050406 Exported Snapshot Timestamp: 12/25 05:04:06 Healthy: true Relationship ID: 078ecbbc-a2e3-11ee-981e-bdd56ead09c8 Source Vserver UUID: 04ee1778-a058-11ee-981e-bdd56ead09c8 Destination Vserver UUID: 5af907bb-a065-11ee-981e-bdd56ead09c8 Current Operation ID: - Transfer Type: - Transfer Error: - Last Transfer Type: update Last Transfer Error: - Last Transfer Error Codes: - Last Transfer Size: 3.07GB Last Transfer Network Compression Ratio: 1:1 Last Transfer Duration: 0:0:50 Last Transfer From: svm:vol1 Last Transfer End Timestamp: 12/25 05:07:47 Unhealthy Reason: - Progress Last Updated: - Relationship Capability: 8.2 and above Lag Time: 0:4:19 Current Transfer Priority: - SMTape Operation: - Destination Volume Node Name: FsxId0ab6f9b00824a187c-01 Identity Preserve Vserver DR: - Number of Successful Updates: 1 Number of Failed Updates: 0 Number of Successful Resyncs: 0 Number of Failed Resyncs: 0 Number of Successful Breaks: 0 Number of Failed Breaks: 0 Total Transfer Bytes: 9057885176 Total Transfer Time in Seconds: 221 Source Volume MSIDs Preserved: - OpMask: ffffffffffffffff Is Auto Expand Enabled: - Percent Complete for Current Status: -
3.07GB転送されました。
Storage Efficiency、ボリューム、aggregate、Snapshotの情報を確認します。
::*> volume efficiency show -volume vol1* -fields state, policy, storage-efficiency-mode, inline-compression, inline-dedupe, compression, data-compaction, auto-adaptive-compression-savings, auto-adaptive-compression-existing-volume, using-auto-adaptive-compression vserver volume state policy compression inline-compression storage-efficiency-mode inline-dedupe data-compaction auto-adaptive-compression-savings using-auto-adaptive-compression auto-adaptive-compression-existing-volume ------- ------ ------- ------ ----------- ------------------ ----------------------- ------------- --------------- --------------------------------- ------------------------------- ----------------------------------------- svm vol1 Enabled auto false true efficient false true true true false svm2 vol1_dst Enabled - false true efficient false true true true false svm2 vol1_dst2 Disabled - false true efficient false true true true false svm3 vol1_dst_dst Enabled auto false true efficient true true true true false 4 entries were displayed. ::*> volume efficiency inactive-data-compression show -volume vol1_dst2 -instance Volume: vol1_dst2 Vserver: svm2 Is Enabled: true Scan Mode: - Progress: IDLE Status: SUCCESS Compression Algorithm: lzopro Failure Reason: - Total Blocks: - Total blocks Processed: - Percentage: - Phase1 L1s Processed: - Phase1 Lns Skipped: - Phase2 Total Blocks: - Phase2 Blocks Processed: - Number of Cold Blocks Encountered: 0 Number of Repacked Blocks: 0 Number of Compression Done Blocks: 0 Number of Vol-Overwrites: 0 Time since Last Inactive Data Compression Scan started(sec): 0 Time since Last Inactive Data Compression Scan ended(sec): 255 Time since Last Successful Inactive Data Compression Scan started(sec): - Time since Last Successful Inactive Data Compression Scan ended(sec): 0 Average time for Cold Data Compression(sec): 0 Tuning Enabled: true Threshold: 14 Threshold Upper Limit: 21 Threshold Lower Limit: 14 Client Read history window: 14 Incompressible Data Percentage: 0% ::*> volume show -volume vol1* -fields available, filesystem-size, total, used, percent-used, performance-tier-inactive-user-data, performance-tier-inactive-user-data-percent, size, dedupe-space-saved, dedupe-space-saved-percent, dedupe-space-shared,logical-used, logical-used-percent,logical-used-by-afs, logical-available vserver volume size available filesystem-size total used percent-used dedupe-space-saved dedupe-space-saved-percent dedupe-space-shared logical-used logical-used-percent logical-available logical-used-by-afs performance-tier-inactive-user-data performance-tier-inactive-user-data-percent ------- ------ ---- --------- --------------- ------- ------ ------------ ------------------ -------------------------- ------------------- ------------ -------------------- ----------------- ------------------- ----------------------------------- ------------------------------------------- svm vol1 16GB 6.85GB 16GB 15.20GB 8.35GB 54% 808.8MB 9% 808.8MB 9.14GB 60% - 9.14GB - - svm2 vol1_dst 7.54GB 1.10GB 7.54GB 7.16GB 6.06GB 84% 4.97GB 45% 3.01GB 11.01GB 154% - 8.05GB 0B 0% svm2 vol1_dst2 10.42GB 1.58GB 10.42GB 9.90GB 8.32GB 84% 774.5MB 8% 8.24GB 9.04GB 91% - 9.04GB 0B 0% svm3 vol1_dst_dst 4.90GB 896.4MB 4.90GB 4.90GB 4.03GB 82% 5.11GB 56% 2.21GB 9.14GB 186% - 9.07GB 0B 0% 4 entries were displayed. ::*> volume show-footprint -volume vol1_dst2 Vserver : svm2 Volume : vol1_dst2 Feature Used Used% -------------------------------- ---------- ----- Volume Data Footprint 8.32GB 1% Footprint in Performance Tier 8.41GB 100% Footprint in FSxFabricpoolObjectStore 0B 0% Volume Guarantee 0B 0% Flexible Volume Metadata 64.09MB 0% Delayed Frees 96.16MB 0% File Operation Metadata 4KB 0% Total Footprint 8.48GB 1% Effective Total Footprint 8.48GB 1% ::*> aggr show-efficiency -instance Name of the Aggregate: aggr1 Node where Aggregate Resides: FsxId0ab6f9b00824a187c-01 Logical Size Used by Volumes, Clones, Snapshot Copies in the Aggregate: 160.2GB Total Physical Used: 19.02GB Total Storage Efficiency Ratio: 8.43:1 Total Data Reduction Logical Used Without Snapshots: 35.01GB Total Data Reduction Physical Used Without Snapshots: 16.35GB Total Data Reduction Efficiency Ratio Without Snapshots: 2.14:1 Total Data Reduction Logical Used without snapshots and flexclones: 35.01GB Total Data Reduction Physical Used without snapshots and flexclones: 16.35GB Total Data Reduction Efficiency Ratio without snapshots and flexclones: 2.14:1 Total Logical Size Used by Volumes, Clones, Snapshot Copies in the FabricPool Performance Tier: 114.3GB Total Physical Used in FabricPool Performance Tier: 15.20GB Total FabricPool Performance Tier Storage Efficiency Ratio: 7.52:1 Total Data Reduction Logical Used without snapshots and flexclones in the FabricPool Performance Tier: 26.50GB Total Data Reduction Physical Used without snapshots and flexclones in the FabricPool Performance Tier: 12.55GB Total FabricPool Performance Tier Data Reduction Efficiency Ratio without snapshots and flexclones: 2.11:1 Logical Space Used for All Volumes: 35.01GB Physical Space Used for All Volumes: 23.38GB Space Saved by Volume Deduplication: 11.63GB Space Saved by Volume Deduplication and pattern detection: 11.63GB Volume Deduplication Savings ratio: 1.50:1 Space Saved by Volume Compression: 0B Volume Compression Savings ratio: 1.00:1 Space Saved by Inline Zero Pattern Detection: 0B Volume Data Reduction SE Ratio: 1.50:1 Logical Space Used by the Aggregate: 24.47GB Physical Space Used by the Aggregate: 19.02GB Space Saved by Aggregate Data Reduction: 5.45GB Aggregate Data Reduction SE Ratio: 1.29:1 Logical Size Used by Snapshot Copies: 125.2GB Physical Size Used by Snapshot Copies: 3.43GB Snapshot Volume Data Reduction Ratio: 36.48:1 Logical Size Used by FlexClone Volumes: 0B Physical Sized Used by FlexClone Volumes: 0B FlexClone Volume Data Reduction Ratio: 1.00:1 Snapshot And FlexClone Volume Data Reduction SE Ratio: 36.48:1 Number of Volumes Offline: 0 Number of SIS Disabled Volumes: 3 Number of SIS Change Log Disabled Volumes: 1 ::*> snapshot show -volume vol1* ---Blocks--- Vserver Volume Snapshot Size Total% Used% -------- -------- ------------------------------------- -------- ------ ----- svm vol1 test.2023-12-22_0533 160KB 0% 0% snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_053528 24.45MB 0% 0% snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_065900 312KB 0% 0% snapmirror.674509fa-a065-11ee-981e-bdd56ead09c8_2148779290.2023-12-22_075507 292KB 0% 0% snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779291.2023-12-25_050406 144KB 0% 0% svm2 vol1_dst snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_015623 1.91GB 25% 38% test.2023-12-22_0533 893.2MB 12% 22% snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_053528 481.9MB 6% 13% snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_065900 79.82MB 1% 2% vol1_dst2 test.2023-12-22_0533 608KB 0% 0% snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_053528 364KB 0% 0% snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_065900 420KB 0% 0% snapmirror.674509fa-a065-11ee-981e-bdd56ead09c8_2148779290.2023-12-22_075507 280KB 0% 0% snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779291.2023-12-25_050406 152KB 0% 0% svm3 vol1_dst_dst snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_065900 24.16MB 0% 1% snapmirror.674509fa-a065-11ee-981e-bdd56ead09c8_2148779290.2023-12-22_074801 41.57MB 1% 1% snapmirror.674509fa-a065-11ee-981e-bdd56ead09c8_2148779290.2023-12-22_075507 4.96MB 0% 0% 17 entries were displayed.
転送先のボリュームでは774.5MBほど重複排除がされています。
また、Space Saved by Aggregate Data Reduction
も変わりありませんでした。
vol1_dst2 でStorage Efficiencyを実行
vol1_dst2
でStorage Efficiencyを実行します。
::*> volume efficiency on -vserver svm2 -volume vol1_dst2 Efficiency for volume "vol1_dst2" of Vserver "svm2" is enabled. ::*> volume efficiency start -vserver svm2 -volume vol1_dst2 -scan-old-data Warning: This operation scans all of the data in volume "vol1_dst2" of Vserver "svm2". It might take a significant time, and degrade performance during that time. Do you want to continue? {y|n}: y The efficiency operation for volume "vol1_dst2" of Vserver "svm2" has started. ::*> volume efficiency show -volume vol1* -fields changelog-usage, changelog-size, logical-data-size, state, progress, last-op-size vserver volume state progress last-op-size changelog-usage changelog-size logical-data-size ------- ------ ------- ----------------- ------------ --------------- -------------- ----------------- svm vol1 Enabled Idle for 69:08:19 1.72GB 5% 8.20MB 9.14GB svm2 vol1_dst Enabled Idle for 00:01:28 12.19GB 0% 120KB 8.08GB svm2 vol1_dst2 Enabled 1757184 KB Scanned 0B 0% 0B 9.08GB svm3 vol1_dst_dst Enabled Idle for 69:43:32 83.62MB 1% 900KB 9.07GB 4 entries were displayed. ::*> volume efficiency show -volume vol1* -fields changelog-usage, changelog-size, logical-data-size, state, progress, last-op-size vserver volume state progress last-op-size changelog-usage changelog-size logical-data-size ------- ------ ------- ----------------- ------------ --------------- -------------- ----------------- svm vol1 Enabled Idle for 69:09:10 1.72GB 5% 8.20MB 9.14GB svm2 vol1_dst Enabled Idle for 00:02:19 12.19GB 0% 120KB 8.08GB svm2 vol1_dst2 Enabled 15507456 KB Scanned 0B 0% 0B 9.08GB svm3 vol1_dst_dst Enabled Idle for 69:44:23 83.62MB 1% 900KB 9.07GB 4 entries were displayed. ::*> volume efficiency show -volume vol1* -fields changelog-usage, changelog-size, logical-data-size, state, progress, last-op-size vserver volume state progress last-op-size changelog-usage changelog-size logical-data-size ------- ------ ------- ----------------- ------------ --------------- -------------- ----------------- svm vol1 Enabled Idle for 69:12:07 1.72GB 5% 8.20MB 9.14GB svm2 vol1_dst Enabled Idle for 00:05:16 12.19GB 0% 120KB 8.08GB svm2 vol1_dst2 Enabled 14701036 KB (91%) Done 0B 0% 3.69MB 9.08GB svm3 vol1_dst_dst Enabled Idle for 69:47:20 83.62MB 1% 900KB 9.07GB 4 entries were displayed. ::*> volume efficiency show -volume vol1* -fields changelog-usage, changelog-size, logical-data-size, state, progress, last-op-size vserver volume state progress last-op-size changelog-usage changelog-size logical-data-size ------- ------ ------- ----------------- ------------ --------------- -------------- ----------------- svm vol1 Enabled Idle for 69:12:51 1.72GB 5% 8.20MB 9.14GB svm2 vol1_dst Enabled Idle for 00:06:00 12.19GB 0% 120KB 8.08GB svm2 vol1_dst2 Enabled Idle for 00:00:16 17.24GB 0% 3.69MB 9.09GB svm3 vol1_dst_dst Enabled Idle for 69:48:04 83.62MB 1% 900KB 9.07GB 4 entries were displayed.
17.24GBに対して処理が行われたようです。
ボリューム、aggregate、Snapshotの情報を確認します。
::*> volume efficiency show -volume vol1* -fields state, policy, storage-efficiency-mode, inline-compression, inline-dedupe, compression, data-compaction, auto-adaptive-compression-savings, auto-adaptive-compression-existing-volume, using-auto-adaptive-compression vserver volume state policy compression inline-compression storage-efficiency-mode inline-dedupe data-compaction auto-adaptive-compression-savings using-auto-adaptive-compression auto-adaptive-compression-existing-volume ------- ------ ------- ------ ----------- ------------------ ----------------------- ------------- --------------- --------------------------------- ------------------------------- ----------------------------------------- svm vol1 Enabled auto false true efficient false true true true false svm2 vol1_dst Enabled - false true efficient false true true true false svm2 vol1_dst2 Enabled - false true efficient false true true true false svm3 vol1_dst_dst Enabled auto false true efficient true true true true false 4 entries were displayed. ::*> volume efficiency inactive-data-compression show -volume vol1_dst2 -instance Volume: vol1_dst2 Vserver: svm2 Is Enabled: true Scan Mode: - Progress: IDLE Status: SUCCESS Compression Algorithm: lzopro Failure Reason: - Total Blocks: - Total blocks Processed: - Percentage: - Phase1 L1s Processed: - Phase1 Lns Skipped: - Phase2 Total Blocks: - Phase2 Blocks Processed: - Number of Cold Blocks Encountered: 0 Number of Repacked Blocks: 0 Number of Compression Done Blocks: 0 Number of Vol-Overwrites: 0 Time since Last Inactive Data Compression Scan started(sec): 0 Time since Last Inactive Data Compression Scan ended(sec): 929 Time since Last Successful Inactive Data Compression Scan started(sec): - Time since Last Successful Inactive Data Compression Scan ended(sec): 0 Average time for Cold Data Compression(sec): 0 Tuning Enabled: true Threshold: 14 Threshold Upper Limit: 21 Threshold Lower Limit: 14 Client Read history window: 14 Incompressible Data Percentage: 0% ::*> volume show -volume vol1* -fields available, filesystem-size, total, used, percent-used, performance-tier-inactive-user-data, performance-tier-inactive-user-data-percent, size, dedupe-space-saved, dedupe-space-saved-percent, dedupe-space-shared,logical-used, logical-used-percent,logical-used-by-afs, logical-available vserver volume size available filesystem-size total used percent-used dedupe-space-saved dedupe-space-saved-percent dedupe-space-shared logical-used logical-used-percent logical-available logical-used-by-afs performance-tier-inactive-user-data performance-tier-inactive-user-data-percent ------- ------ ---- --------- --------------- ------- ------ ------------ ------------------ -------------------------- ------------------- ------------ -------------------- ----------------- ------------------- ----------------------------------- ------------------------------------------- svm vol1 16GB 6.85GB 16GB 15.20GB 8.35GB 54% 808.8MB 9% 808.8MB 9.14GB 60% - 9.14GB - - svm2 vol1_dst 7.54GB 1.10GB 7.54GB 7.16GB 6.06GB 84% 4.98GB 45% 3.00GB 11.01GB 154% - 8.05GB 0B 0% svm2 vol1_dst2 9.84GB 1.45GB 9.84GB 9.35GB 7.90GB 84% 4.53GB 36% 4.10GB 12.40GB 133% - 9.06GB 0B 0% svm3 vol1_dst_dst 4.90GB 896.4MB 4.90GB 4.90GB 4.03GB 82% 5.11GB 56% 2.21GB 9.14GB 186% - 9.07GB 0B 0% 4 entries were displayed. ::*> volume show-footprint -volume vol1_dst2 Vserver : svm2 Volume : vol1_dst2 Feature Used Used% -------------------------------- ---------- ----- Volume Data Footprint 8.39GB 1% Footprint in Performance Tier 8.46GB 100% Footprint in FSxFabricpoolObjectStore 0B 0% Volume Guarantee 0B 0% Flexible Volume Metadata 64.09MB 0% Deduplication Metadata 12.04MB 0% Deduplication 12.04MB 0% Delayed Frees 69.02MB 0% File Operation Metadata 4KB 0% Total Footprint 8.53GB 1% Effective Total Footprint 8.53GB 1% ::*> aggr show-efficiency -instance Name of the Aggregate: aggr1 Node where Aggregate Resides: FsxId0ab6f9b00824a187c-01 Logical Size Used by Volumes, Clones, Snapshot Copies in the Aggregate: 160.2GB Total Physical Used: 19.20GB Total Storage Efficiency Ratio: 8.35:1 Total Data Reduction Logical Used Without Snapshots: 35.02GB Total Data Reduction Physical Used Without Snapshots: 13.54GB Total Data Reduction Efficiency Ratio Without Snapshots: 2.59:1 Total Data Reduction Logical Used without snapshots and flexclones: 35.02GB Total Data Reduction Physical Used without snapshots and flexclones: 13.54GB Total Data Reduction Efficiency Ratio without snapshots and flexclones: 2.59:1 Total Logical Size Used by Volumes, Clones, Snapshot Copies in the FabricPool Performance Tier: 114.3GB Total Physical Used in FabricPool Performance Tier: 15.41GB Total FabricPool Performance Tier Storage Efficiency Ratio: 7.42:1 Total Data Reduction Logical Used without snapshots and flexclones in the FabricPool Performance Tier: 26.51GB Total Data Reduction Physical Used without snapshots and flexclones in the FabricPool Performance Tier: 9.76GB Total FabricPool Performance Tier Data Reduction Efficiency Ratio without snapshots and flexclones: 2.72:1 Logical Space Used for All Volumes: 35.02GB Physical Space Used for All Volumes: 19.61GB Space Saved by Volume Deduplication: 15.41GB Space Saved by Volume Deduplication and pattern detection: 15.41GB Volume Deduplication Savings ratio: 1.79:1 Space Saved by Volume Compression: 0B Volume Compression Savings ratio: 1.00:1 Space Saved by Inline Zero Pattern Detection: 0B Volume Data Reduction SE Ratio: 1.79:1 Logical Space Used by the Aggregate: 24.65GB Physical Space Used by the Aggregate: 19.20GB Space Saved by Aggregate Data Reduction: 5.45GB Aggregate Data Reduction SE Ratio: 1.28:1 Logical Size Used by Snapshot Copies: 125.2GB Physical Size Used by Snapshot Copies: 7.27GB Snapshot Volume Data Reduction Ratio: 17.23:1 Logical Size Used by FlexClone Volumes: 0B Physical Sized Used by FlexClone Volumes: 0B FlexClone Volume Data Reduction Ratio: 1.00:1 Snapshot And FlexClone Volume Data Reduction SE Ratio: 17.23:1 Number of Volumes Offline: 0 Number of SIS Disabled Volumes: 3 Number of SIS Change Log Disabled Volumes: 0 ::*> snapshot show -volume vol1* ---Blocks--- Vserver Volume Snapshot Size Total% Used% -------- -------- ------------------------------------- -------- ------ ----- svm vol1 test.2023-12-22_0533 160KB 0% 0% snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_053528 24.45MB 0% 0% snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_065900 312KB 0% 0% snapmirror.674509fa-a065-11ee-981e-bdd56ead09c8_2148779290.2023-12-22_075507 292KB 0% 0% snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779291.2023-12-25_050406 144KB 0% 0% svm2 vol1_dst snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_015623 1.91GB 25% 38% test.2023-12-22_0533 893.2MB 12% 22% snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_053528 481.9MB 6% 13% snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_065900 82.71MB 1% 3% vol1_dst2 test.2023-12-22_0533 608KB 0% 0% snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_053528 364KB 0% 0% snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_065900 420KB 0% 0% snapmirror.674509fa-a065-11ee-981e-bdd56ead09c8_2148779290.2023-12-22_075507 280KB 0% 0% snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779291.2023-12-25_050406 3.83GB 39% 46% svm3 vol1_dst_dst snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_065900 24.16MB 0% 1% snapmirror.674509fa-a065-11ee-981e-bdd56ead09c8_2148779290.2023-12-22_074801 41.57MB 1% 1% snapmirror.674509fa-a065-11ee-981e-bdd56ead09c8_2148779290.2023-12-22_075507 4.96MB 0% 0% 17 entries were displayed.
4.53GB重複排除されたようです。
vol1_dst2 でInactive data compressionを実行
vol1_dst2
でInactive data compressionを実行します。
::*> volume efficiency inactive-data-compression start -vserver svm2 -volume vol1_dst2 -inactive-days 0 Inactive data compression scan started on volume "vol1_dst2" in Vserver "svm2" ::*> volume efficiency inactive-data-compression show -volume vol1_dst2 -instance Volume: vol1_dst2 Vserver: svm2 Is Enabled: true Scan Mode: default Progress: RUNNING Status: SUCCESS Compression Algorithm: lzopro Failure Reason: - Total Blocks: - Total blocks Processed: - Percentage: 0% Phase1 L1s Processed: 498 Phase1 Lns Skipped: L1: 0 L2: 0 L3: 0 L4: 0 L5: 0 L6: 0 L7: 0 Phase2 Total Blocks: 0 Phase2 Blocks Processed: 0 Number of Cold Blocks Encountered: 100576 Number of Repacked Blocks: 0 Number of Compression Done Blocks: 3592 Number of Vol-Overwrites: 0 Time since Last Inactive Data Compression Scan started(sec): 0 Time since Last Inactive Data Compression Scan ended(sec): 0 Time since Last Successful Inactive Data Compression Scan started(sec): - Time since Last Successful Inactive Data Compression Scan ended(sec): 0 Average time for Cold Data Compression(sec): 0 Tuning Enabled: true Threshold: 14 Threshold Upper Limit: 21 Threshold Lower Limit: 14 Client Read history window: 14 Incompressible Data Percentage: 0% ::*> volume efficiency inactive-data-compression show -volume vol1_dst2 -instance Volume: vol1_dst2 Vserver: svm2 Is Enabled: true Scan Mode: default Progress: RUNNING Status: SUCCESS Compression Algorithm: lzopro Failure Reason: - Total Blocks: - Total blocks Processed: - Percentage: 83% Phase1 L1s Processed: 9252 Phase1 Lns Skipped: L1: 0 L2: 0 L3: 0 L4: 0 L5: 0 L6: 0 L7: 0 Phase2 Total Blocks: 3281776 Phase2 Blocks Processed: 2735104 Number of Cold Blocks Encountered: 1275784 Number of Repacked Blocks: 0 Number of Compression Done Blocks: 615760 Number of Vol-Overwrites: 0 Time since Last Inactive Data Compression Scan started(sec): 605 Time since Last Inactive Data Compression Scan ended(sec): 515 Time since Last Successful Inactive Data Compression Scan started(sec): - Time since Last Successful Inactive Data Compression Scan ended(sec): 515 Average time for Cold Data Compression(sec): 0 Tuning Enabled: true Threshold: 14 Threshold Upper Limit: 21 Threshold Lower Limit: 14 Client Read history window: 14 Incompressible Data Percentage: 0% ::*> volume efficiency inactive-data-compression show -volume vol1_dst2 -instance Volume: vol1_dst2 Vserver: svm2 Is Enabled: true Scan Mode: - Progress: IDLE Status: SUCCESS Compression Algorithm: lzopro Failure Reason: - Total Blocks: - Total blocks Processed: - Percentage: - Phase1 L1s Processed: - Phase1 Lns Skipped: - Phase2 Total Blocks: - Phase2 Blocks Processed: - Number of Cold Blocks Encountered: 2297288 Number of Repacked Blocks: 0 Number of Compression Done Blocks: 1002592 Number of Vol-Overwrites: 0 Time since Last Inactive Data Compression Scan started(sec): 550 Time since Last Inactive Data Compression Scan ended(sec): 27 Time since Last Successful Inactive Data Compression Scan started(sec): - Time since Last Successful Inactive Data Compression Scan ended(sec): 27 Average time for Cold Data Compression(sec): 523 Tuning Enabled: true Threshold: 14 Threshold Upper Limit: 21 Threshold Lower Limit: 14 Client Read history window: 14 Incompressible Data Percentage: 50%
Storage Efficiency、ボリューム、aggregate、Snapshotの情報を確認します。
::*> volume efficiency show -volume vol1* -fields state, policy, storage-efficiency-mode, inline-compression, inline-dedupe, compression, data-compaction, auto-adaptive-compression-savings, auto-adaptive-compression-existing-volume, using-auto-adaptive-compression vserver volume state policy compression inline-compression storage-efficiency-mode inline-dedupe data-compaction auto-adaptive-compression-savings using-auto-adaptive-compression auto-adaptive-compression-existing-volume ------- ------ ------- ------ ----------- ------------------ ----------------------- ------------- --------------- --------------------------------- ------------------------------- ----------------------------------------- svm vol1 Enabled auto false true efficient false true true true false svm2 vol1_dst Enabled - false true efficient false true true true false svm2 vol1_dst2 Enabled - false true efficient false true true true false svm3 vol1_dst_dst Enabled auto false true efficient true true true true false 4 entries were displayed. ::*> volume show -volume vol1* -fields available, filesystem-size, total, used, percent-used, performance-tier-inactive-user-data, performance-tier-inactive-user-data-percent, size, dedupe-space-saved, dedupe-space-saved-percent, dedupe-space-shared,logical-used, logical-used-percent,logical-used-by-afs, logical-available vserver volume size available filesystem-size total used percent-used dedupe-space-saved dedupe-space-saved-percent dedupe-space-shared logical-used logical-used-percent logical-available logical-used-by-afs performance-tier-inactive-user-data performance-tier-inactive-user-data-percent ------- ------ ---- --------- --------------- ------- ------ ------------ ------------------ -------------------------- ------------------- ------------ -------------------- ----------------- ------------------- ----------------------------------- ------------------------------------------- svm vol1 16GB 6.85GB 16GB 15.20GB 8.35GB 54% 808.8MB 9% 808.8MB 9.14GB 60% - 9.14GB - - svm2 vol1_dst 7.54GB 1.10GB 7.54GB 7.16GB 6.06GB 84% 4.98GB 45% 3.00GB 11.01GB 154% - 8.05GB 0B 0% svm2 vol1_dst2 9.84GB 1.45GB 9.84GB 9.35GB 7.90GB 84% 4.53GB 36% 4.10GB 12.40GB 133% - 9.06GB 0B 0% svm3 vol1_dst_dst 4.90GB 896.4MB 4.90GB 4.90GB 4.03GB 82% 5.11GB 56% 2.21GB 9.14GB 186% - 9.07GB 0B 0% 4 entries were displayed. ::*> volume show-footprint -volume vol1_dst2 Vserver : svm2 Volume : vol1_dst2 Feature Used Used% -------------------------------- ---------- ----- Volume Data Footprint 8.39GB 1% Footprint in Performance Tier 8.47GB 100% Footprint in FSxFabricpoolObjectStore 0B 0% Volume Guarantee 0B 0% Flexible Volume Metadata 64.09MB 0% Deduplication Metadata 12.04MB 0% Deduplication 12.04MB 0% Delayed Frees 80.61MB 0% File Operation Metadata 4KB 0% Total Footprint 8.54GB 1% Footprint Data Reduction 3.87GB 0% Auto Adaptive Compression 3.87GB 0% Effective Total Footprint 4.67GB 1% ::*> aggr show-efficiency -instance Name of the Aggregate: aggr1 Node where Aggregate Resides: FsxId0ab6f9b00824a187c-01 Logical Size Used by Volumes, Clones, Snapshot Copies in the Aggregate: 160.1GB Total Physical Used: 19.06GB Total Storage Efficiency Ratio: 8.40:1 Total Data Reduction Logical Used Without Snapshots: 34.90GB Total Data Reduction Physical Used Without Snapshots: 13.92GB Total Data Reduction Efficiency Ratio Without Snapshots: 2.51:1 Total Data Reduction Logical Used without snapshots and flexclones: 34.90GB Total Data Reduction Physical Used without snapshots and flexclones: 13.92GB Total Data Reduction Efficiency Ratio without snapshots and flexclones: 2.51:1 Total Logical Size Used by Volumes, Clones, Snapshot Copies in the FabricPool Performance Tier: 114.3GB Total Physical Used in FabricPool Performance Tier: 15.38GB Total FabricPool Performance Tier Storage Efficiency Ratio: 7.44:1 Total Data Reduction Logical Used without snapshots and flexclones in the FabricPool Performance Tier: 26.51GB Total Data Reduction Physical Used without snapshots and flexclones in the FabricPool Performance Tier: 10.26GB Total FabricPool Performance Tier Data Reduction Efficiency Ratio without snapshots and flexclones: 2.59:1 Logical Space Used for All Volumes: 34.90GB Physical Space Used for All Volumes: 19.49GB Space Saved by Volume Deduplication: 15.41GB Space Saved by Volume Deduplication and pattern detection: 15.41GB Volume Deduplication Savings ratio: 1.79:1 Space Saved by Volume Compression: 0B Volume Compression Savings ratio: 1.00:1 Space Saved by Inline Zero Pattern Detection: 0B Volume Data Reduction SE Ratio: 1.79:1 Logical Space Used by the Aggregate: 26.96GB Physical Space Used by the Aggregate: 19.06GB Space Saved by Aggregate Data Reduction: 7.90GB Aggregate Data Reduction SE Ratio: 1.41:1 Logical Size Used by Snapshot Copies: 125.2GB Physical Size Used by Snapshot Copies: 7.27GB Snapshot Volume Data Reduction Ratio: 17.23:1 Logical Size Used by FlexClone Volumes: 0B Physical Sized Used by FlexClone Volumes: 0B FlexClone Volume Data Reduction Ratio: 1.00:1 Snapshot And FlexClone Volume Data Reduction SE Ratio: 17.23:1 Number of Volumes Offline: 0 Number of SIS Disabled Volumes: 3 Number of SIS Change Log Disabled Volumes: 0 ::*> snapshot show -volume vol1* ---Blocks--- Vserver Volume Snapshot Size Total% Used% -------- -------- ------------------------------------- -------- ------ ----- svm vol1 test.2023-12-22_0533 160KB 0% 0% snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_053528 24.45MB 0% 0% snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_065900 312KB 0% 0% snapmirror.674509fa-a065-11ee-981e-bdd56ead09c8_2148779290.2023-12-22_075507 292KB 0% 0% snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779291.2023-12-25_050406 144KB 0% 0% svm2 vol1_dst snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_015623 1.91GB 25% 38% test.2023-12-22_0533 893.2MB 12% 22% snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_053528 481.9MB 6% 13% snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_065900 82.71MB 1% 3% vol1_dst2 test.2023-12-22_0533 608KB 0% 0% snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_053528 364KB 0% 0% snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_065900 420KB 0% 0% snapmirror.674509fa-a065-11ee-981e-bdd56ead09c8_2148779290.2023-12-22_075507 280KB 0% 0% snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779291.2023-12-25_050406 3.83GB 39% 46% svm3 vol1_dst_dst snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_065900 24.16MB 0% 1% snapmirror.674509fa-a065-11ee-981e-bdd56ead09c8_2148779290.2023-12-22_074801 41.57MB 1% 1% snapmirror.674509fa-a065-11ee-981e-bdd56ead09c8_2148779290.2023-12-22_075507 4.96MB 0% 0% 17 entries were displayed.
Space Saved by Aggregate Data Reduction
が5.45GBから7.90GBと2.45GBほど増加しています。
また、Auto Adaptive Compression
は3.87GBと表示されるようになりました。
svm3 のボリュームの作成
svm3にボリュームvol1_dst2_dst
を作成します。
::*> volume create -vserver svm3 -volume vol1_dst2_dst -aggregate aggr1 -state online -type DP -size 4GB -tiering-policy none [Job 113] Job succeeded: Successful ::*> volume show -volume vol1* -fields type, autosize-mode, max-autosize vserver volume max-autosize autosize-mode type ------- ------ ------------ ------------- ---- svm vol1 19.20GB off RW svm2 vol1_dst 100TB grow_shrink DP svm2 vol1_dst2 100TB grow_shrink DP svm3 vol1_dst2_dst 100TB grow_shrink DP svm3 vol1_dst_dst 100TB grow_shrink RW 5 entries were displayed.
svm2 と svm3 のボリューム間のSnapMirror initialize
svm2 と svm3 のボリューム間のSnapMirror relationshipのinitializeを行います。
まず、SnapMirror relasionshipの作成をします。
::*> snapmirror create -source-path svm2:vol1_dst2 -destination-vserver svm3 -destination-volume vol1_dst2_dst -policy MirrorAllSnapshots Operation succeeded: snapmirror create for the relationship with destination "svm3:vol1_dst2_dst".
次にinitializeを行います。
::*> snapmirror initialize -destination-path svm3:vol1_dst2_dst -source-path svm2:vol1_dst2 Operation is queued: snapmirror initialize of destination "svm3:vol1_dst2_dst". ::*> snapmirror show Progress Source Destination Mirror Relationship Total Last Path Type Path State Status Progress Healthy Updated ----------- ---- ------------ ------- -------------- --------- ------- -------- svm:vol1 XDP svm2:vol1_dst Snapmirrored Idle - true - svm2:vol1_dst2 Snapmirrored Idle - true - svm3:vol1_dst_dst Broken-off Idle - true - svm2:vol1_dst2 XDP svm3:vol1_dst2_dst Uninitialized Transferring 0B true 12/25 05:45:42 4 entries were displayed. ::*> snapmirror show Progress Source Destination Mirror Relationship Total Last Path Type Path State Status Progress Healthy Updated ----------- ---- ------------ ------- -------------- --------- ------- -------- svm:vol1 XDP svm2:vol1_dst Snapmirrored Idle - true - svm2:vol1_dst2 Snapmirrored Idle - true - svm3:vol1_dst_dst Broken-off Idle - true - svm2:vol1_dst2 XDP svm3:vol1_dst2_dst Uninitialized Transferring 1.33GB true 12/25 05:45:50 4 entries were displayed. ::*> snapmirror show Progress Source Destination Mirror Relationship Total Last Path Type Path State Status Progress Healthy Updated ----------- ---- ------------ ------- -------------- --------- ------- -------- svm:vol1 XDP svm2:vol1_dst Snapmirrored Idle - true - svm2:vol1_dst2 Snapmirrored Idle - true - svm3:vol1_dst_dst Broken-off Idle - true - svm2:vol1_dst2 XDP svm3:vol1_dst2_dst Snapmirrored Transferring 1.31GB true 12/25 05:46:41 4 entries were displayed. ::*> snapmirror show Progress Source Destination Mirror Relationship Total Last Path Type Path State Status Progress Healthy Updated ----------- ---- ------------ ------- -------------- --------- ------- -------- svm:vol1 XDP svm2:vol1_dst Snapmirrored Idle - true - svm2:vol1_dst2 Snapmirrored Idle - true - svm3:vol1_dst_dst Broken-off Idle - true - svm2:vol1_dst2 XDP svm3:vol1_dst2_dst Snapmirrored Idle - true - 4 entries were displayed. ::*> snapmirror show -destination-path svm3:vol1_dst2_dst Source Path: svm2:vol1_dst2 Source Cluster: - Source Vserver: svm2 Source Volume: vol1_dst2 Destination Path: svm3:vol1_dst2_dst Destination Cluster: - Destination Vserver: svm3 Destination Volume: vol1_dst2_dst Relationship Type: XDP Relationship Group Type: none Managing Vserver: svm3 SnapMirror Schedule: - SnapMirror Policy Type: async-mirror SnapMirror Policy: MirrorAllSnapshots Tries Limit: - Throttle (KB/sec): unlimited Consistency Group Item Mappings: - Current Transfer Throttle (KB/sec): - Mirror State: Snapmirrored Relationship Status: Idle File Restore File Count: - File Restore File List: - Transfer Snapshot: - Snapshot Progress: - Total Progress: - Network Compression Ratio: - Snapshot Checkpoint: - Newest Snapshot: snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779291.2023-12-25_050406 Newest Snapshot Timestamp: 12/25 05:04:06 Exported Snapshot: snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779291.2023-12-25_050406 Exported Snapshot Timestamp: 12/25 05:04:06 Healthy: true Relationship ID: cb1dc5b5-a2e8-11ee-981e-bdd56ead09c8 Source Vserver UUID: 5af907bb-a065-11ee-981e-bdd56ead09c8 Destination Vserver UUID: 674509fa-a065-11ee-981e-bdd56ead09c8 Current Operation ID: - Transfer Type: - Transfer Error: - Last Transfer Type: update Last Transfer Error: - Last Transfer Error Codes: - Last Transfer Size: 1.59GB Last Transfer Network Compression Ratio: 1:1 Last Transfer Duration: 0:0:38 Last Transfer From: svm2:vol1_dst2 Last Transfer End Timestamp: 12/25 05:46:58 Unhealthy Reason: - Progress Last Updated: - Relationship Capability: 8.2 and above Lag Time: 0:43:37 Current Transfer Priority: - SMTape Operation: - Destination Volume Node Name: FsxId0ab6f9b00824a187c-01 Identity Preserve Vserver DR: - Number of Successful Updates: 1 Number of Failed Updates: 0 Number of Successful Resyncs: 0 Number of Failed Resyncs: 0 Number of Successful Breaks: 0 Number of Failed Breaks: 0 Total Transfer Bytes: 5891022441 Total Transfer Time in Seconds: 76 Source Volume MSIDs Preserved: - OpMask: ffffffffffffffff Is Auto Expand Enabled: - Percent Complete for Current Status: -
Storage Efficiency、ボリューム、aggregate、Snapshotの情報を確認します。
::*> volume efficiency show -volume vol1* -fields state, policy, storage-efficiency-mode, inline-compression, inline-dedupe, compression, data-compaction, auto-adaptive-compression-savings, auto-adaptive-compression-existing-volume, using-auto-adaptive-compression vserver volume state policy compression inline-compression storage-efficiency-mode inline-dedupe data-compaction auto-adaptive-compression-savings using-auto-adaptive-compression auto-adaptive-compression-existing-volume ------- ------ ------- ------ ----------- ------------------ ----------------------- ------------- --------------- --------------------------------- ------------------------------- ----------------------------------------- svm vol1 Enabled auto false true efficient false true true true false svm2 vol1_dst Enabled - false true efficient false true true true false svm2 vol1_dst2 Enabled - false true efficient false true true true false svm3 vol1_dst2_dst Disabled - false true efficient false true true true false svm3 vol1_dst_dst Enabled auto false true efficient true true true true false 5 entries were displayed. ::*> volume efficiency inactive-data-compression show -volume vol1_dst2_dst -instance Volume: vol1_dst2_dst Vserver: svm3 Is Enabled: true Scan Mode: - Progress: IDLE Status: SUCCESS Compression Algorithm: lzopro Failure Reason: - Total Blocks: - Total blocks Processed: - Percentage: - Phase1 L1s Processed: - Phase1 Lns Skipped: - Phase2 Total Blocks: - Phase2 Blocks Processed: - Number of Cold Blocks Encountered: 0 Number of Repacked Blocks: 0 Number of Compression Done Blocks: 0 Number of Vol-Overwrites: 0 Time since Last Inactive Data Compression Scan started(sec): 0 Time since Last Inactive Data Compression Scan ended(sec): 134 Time since Last Successful Inactive Data Compression Scan started(sec): - Time since Last Successful Inactive Data Compression Scan ended(sec): 0 Average time for Cold Data Compression(sec): 0 Tuning Enabled: true Threshold: 14 Threshold Upper Limit: 21 Threshold Lower Limit: 14 Client Read history window: 14 Incompressible Data Percentage: 0% ::*> volume show -volume vol1* -fields available, filesystem-size, total, used, percent-used, performance-tier-inactive-user-data, performance-tier-inactive-user-data-percent, size, dedupe-space-saved, dedupe-space-saved-percent, dedupe-space-shared,logical-used, logical-used-percent,logical-used-by-afs, logical-available vserver volume size available filesystem-size total used percent-used dedupe-space-saved dedupe-space-saved-percent dedupe-space-shared logical-used logical-used-percent logical-available logical-used-by-afs performance-tier-inactive-user-data performance-tier-inactive-user-data-percent ------- ------ ---- --------- --------------- ------- ------ ------------ ------------------ -------------------------- ------------------- ------------ -------------------- ----------------- ------------------- ----------------------------------- ------------------------------------------- svm vol1 16GB 6.85GB 16GB 15.20GB 8.35GB 54% 808.8MB 9% 808.8MB 9.14GB 60% - 9.14GB - - svm2 vol1_dst 7.54GB 1.10GB 7.54GB 7.16GB 6.06GB 84% 4.98GB 45% 3.00GB 11.01GB 154% - 8.05GB 0B 0% svm2 vol1_dst2 9.84GB 1.45GB 9.84GB 9.35GB 7.90GB 84% 4.53GB 36% 4.10GB 12.40GB 133% - 9.06GB 0B 0% svm3 vol1_dst2_dst 9.88GB 1.55GB 9.88GB 9.88GB 8.32GB 84% 774.5MB 8% 8.24GB 9.04GB 92% - 9.04GB 0B 0% svm3 vol1_dst_dst 4.90GB 896.4MB 4.90GB 4.90GB 4.03GB 82% 5.11GB 56% 2.21GB 9.14GB 186% - 9.07GB 0B 0% 5 entries were displayed. ::*> volume show-footprint -volume vol1_dst2_dst Vserver : svm3 Volume : vol1_dst2_dst Feature Used Used% -------------------------------- ---------- ----- Volume Data Footprint 8.32GB 1% Footprint in Performance Tier 8.42GB 100% Footprint in FSxFabricpoolObjectStore 0B 0% Volume Guarantee 0B 0% Flexible Volume Metadata 61.31MB 0% Delayed Frees 101.0MB 0% File Operation Metadata 4KB 0% Total Footprint 8.48GB 1% Footprint Data Reduction 3.71GB 0% Auto Adaptive Compression 3.71GB 0% Effective Total Footprint 4.77GB 1% ::*> aggr show-efficiency -instance Name of the Aggregate: aggr1 Node where Aggregate Resides: FsxId0ab6f9b00824a187c-01 Logical Size Used by Volumes, Clones, Snapshot Copies in the Aggregate: 207.4GB Total Physical Used: 23.04GB Total Storage Efficiency Ratio: 9.00:1 Total Data Reduction Logical Used Without Snapshots: 43.97GB Total Data Reduction Physical Used Without Snapshots: 17.99GB Total Data Reduction Efficiency Ratio Without Snapshots: 2.44:1 Total Data Reduction Logical Used without snapshots and flexclones: 43.97GB Total Data Reduction Physical Used without snapshots and flexclones: 17.99GB Total Data Reduction Efficiency Ratio without snapshots and flexclones: 2.44:1 Total Logical Size Used by Volumes, Clones, Snapshot Copies in the FabricPool Performance Tier: 161.6GB Total Physical Used in FabricPool Performance Tier: 19.38GB Total FabricPool Performance Tier Storage Efficiency Ratio: 8.34:1 Total Data Reduction Logical Used without snapshots and flexclones in the FabricPool Performance Tier: 35.59GB Total Data Reduction Physical Used without snapshots and flexclones in the FabricPool Performance Tier: 14.35GB Total FabricPool Performance Tier Data Reduction Efficiency Ratio without snapshots and flexclones: 2.48:1 Logical Space Used for All Volumes: 43.97GB Physical Space Used for All Volumes: 27.80GB Space Saved by Volume Deduplication: 16.17GB Space Saved by Volume Deduplication and pattern detection: 16.17GB Volume Deduplication Savings ratio: 1.58:1 Space Saved by Volume Compression: 0B Volume Compression Savings ratio: 1.00:1 Space Saved by Inline Zero Pattern Detection: 0B Volume Data Reduction SE Ratio: 1.58:1 Logical Space Used by the Aggregate: 33.19GB Physical Space Used by the Aggregate: 23.04GB Space Saved by Aggregate Data Reduction: 10.15GB Aggregate Data Reduction SE Ratio: 1.44:1 Logical Size Used by Snapshot Copies: 163.4GB Physical Size Used by Snapshot Copies: 7.27GB Snapshot Volume Data Reduction Ratio: 22.48:1 Logical Size Used by FlexClone Volumes: 0B Physical Sized Used by FlexClone Volumes: 0B FlexClone Volume Data Reduction Ratio: 1.00:1 Snapshot And FlexClone Volume Data Reduction SE Ratio: 22.48:1 Number of Volumes Offline: 0 Number of SIS Disabled Volumes: 3 Number of SIS Change Log Disabled Volumes: 1 ::*> snapshot show -volume vol1* ---Blocks--- Vserver Volume Snapshot Size Total% Used% -------- -------- ------------------------------------- -------- ------ ----- svm vol1 test.2023-12-22_0533 160KB 0% 0% snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_053528 24.45MB 0% 0% snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_065900 312KB 0% 0% snapmirror.674509fa-a065-11ee-981e-bdd56ead09c8_2148779290.2023-12-22_075507 292KB 0% 0% snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779291.2023-12-25_050406 144KB 0% 0% svm2 vol1_dst snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_015623 1.91GB 25% 38% test.2023-12-22_0533 893.2MB 12% 22% snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_053528 481.9MB 6% 13% snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_065900 82.71MB 1% 3% vol1_dst2 test.2023-12-22_0533 608KB 0% 0% snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_053528 364KB 0% 0% snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_065900 420KB 0% 0% snapmirror.674509fa-a065-11ee-981e-bdd56ead09c8_2148779290.2023-12-22_075507 280KB 0% 0% snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779291.2023-12-25_050406 3.83GB 39% 46% svm3 vol1_dst2_dst test.2023-12-22_0533 536KB 0% 0% snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_053528 380KB 0% 0% snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_065900 380KB 0% 0% snapmirror.674509fa-a065-11ee-981e-bdd56ead09c8_2148779290.2023-12-22_075507 300KB 0% 0% snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779291.2023-12-25_050406 152KB 0% 0% vol1_dst_dst snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_065900 24.16MB 0% 1% snapmirror.674509fa-a065-11ee-981e-bdd56ead09c8_2148779290.2023-12-22_074801 41.57MB 1% 1% snapmirror.674509fa-a065-11ee-981e-bdd56ead09c8_2148779290.2023-12-22_075507 4.96MB 0% 0% 22 entries were displayed.
vol1_dst2
で実行したStorage Efficiencyの重複排除はvol1_dst2_dst
に引き継がれていませんが、Space Saved by Aggregate Data Reduction
が7.90GBから10.15GBと2.25GBほど増加しています。
また、Auto Adaptive Compression
は3.71GBと表示されています。
そのため、aggregateレイヤーのデータ削減効果はしっかりと維持できていそうです。
svm2 と svm3 のボリューム間のSnapMirror update
svm2 と svm3 のボリューム間のSnapMirrorの差分転送を行います。
::*> snapmirror update -destination-path svm3:vol1_dst2_dst Operation is queued: snapmirror update of destination "svm3:vol1_dst2_dst". ::*> snapmirror show Progress Source Destination Mirror Relationship Total Last Path Type Path State Status Progress Healthy Updated ----------- ---- ------------ ------- -------------- --------- ------- -------- svm:vol1 XDP svm2:vol1_dst Snapmirrored Idle - true - svm2:vol1_dst2 Snapmirrored Idle - true - svm3:vol1_dst_dst Broken-off Idle - true - svm2:vol1_dst2 XDP svm3:vol1_dst2_dst Snapmirrored Idle - true - 4 entries were displayed. ::*> snapmirror show -destination-path svm3:vol1_dst2_dst Source Path: svm2:vol1_dst2 Source Cluster: - Source Vserver: svm2 Source Volume: vol1_dst2 Destination Path: svm3:vol1_dst2_dst Destination Cluster: - Destination Vserver: svm3 Destination Volume: vol1_dst2_dst Relationship Type: XDP Relationship Group Type: none Managing Vserver: svm3 SnapMirror Schedule: - SnapMirror Policy Type: async-mirror SnapMirror Policy: MirrorAllSnapshots Tries Limit: - Throttle (KB/sec): unlimited Consistency Group Item Mappings: - Current Transfer Throttle (KB/sec): - Mirror State: Snapmirrored Relationship Status: Idle File Restore File Count: - File Restore File List: - Transfer Snapshot: - Snapshot Progress: - Total Progress: - Network Compression Ratio: - Snapshot Checkpoint: - Newest Snapshot: snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779291.2023-12-25_050406 Newest Snapshot Timestamp: 12/25 05:04:06 Exported Snapshot: snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779291.2023-12-25_050406 Exported Snapshot Timestamp: 12/25 05:04:06 Healthy: true Relationship ID: cb1dc5b5-a2e8-11ee-981e-bdd56ead09c8 Source Vserver UUID: 5af907bb-a065-11ee-981e-bdd56ead09c8 Destination Vserver UUID: 674509fa-a065-11ee-981e-bdd56ead09c8 Current Operation ID: - Transfer Type: - Transfer Error: - Last Transfer Type: update Last Transfer Error: - Last Transfer Error Codes: - Last Transfer Size: 0B Last Transfer Network Compression Ratio: 1:1 Last Transfer Duration: 0:0:0 Last Transfer From: svm2:vol1_dst2 Last Transfer End Timestamp: 12/25 05:52:44 Unhealthy Reason: - Progress Last Updated: - Relationship Capability: 8.2 and above Lag Time: 0:49:1 Current Transfer Priority: - SMTape Operation: - Destination Volume Node Name: FsxId0ab6f9b00824a187c-01 Identity Preserve Vserver DR: - Number of Successful Updates: 2 Number of Failed Updates: 0 Number of Successful Resyncs: 0 Number of Failed Resyncs: 0 Number of Successful Breaks: 0 Number of Failed Breaks: 0 Total Transfer Bytes: 5891022441 Total Transfer Time in Seconds: 76 Source Volume MSIDs Preserved: - OpMask: ffffffffffffffff Is Auto Expand Enabled: - Percent Complete for Current Status: -
転送量は0Bです。
Storage Efficiency、ボリューム、aggregate、Snapshotの情報を確認します。
::*> volume efficiency show -volume vol1* -fields state, policy, storage-efficiency-mode, inline-compression, inline-dedupe, compression, data-compaction, auto-adaptive-compression-savings, auto-adaptive-compression-existing-volume, using-auto-adaptive-compression vserver volume state policy compression inline-compression storage-efficiency-mode inline-dedupe data-compaction auto-adaptive-compression-savings using-auto-adaptive-compression auto-adaptive-compression-existing-volume ------- ------ ------- ------ ----------- ------------------ ----------------------- ------------- --------------- --------------------------------- ------------------------------- ----------------------------------------- svm vol1 Enabled auto false true efficient false true true true false svm2 vol1_dst Enabled - false true efficient false true true true false svm2 vol1_dst2 Enabled - false true efficient false true true true false svm3 vol1_dst2_dst Disabled - false true efficient false true true true false svm3 vol1_dst_dst Enabled auto false true efficient true true true true false 5 entries were displayed. ::*> volume efficiency inactive-data-compression show -volume vol1_dst2_dst -instance Volume: vol1_dst2_dst Vserver: svm3 Is Enabled: true Scan Mode: - Progress: IDLE Status: SUCCESS Compression Algorithm: lzopro Failure Reason: - Total Blocks: - Total blocks Processed: - Percentage: - Phase1 L1s Processed: - Phase1 Lns Skipped: - Phase2 Total Blocks: - Phase2 Blocks Processed: - Number of Cold Blocks Encountered: 0 Number of Repacked Blocks: 0 Number of Compression Done Blocks: 0 Number of Vol-Overwrites: 0 Time since Last Inactive Data Compression Scan started(sec): 0 Time since Last Inactive Data Compression Scan ended(sec): 797 Time since Last Successful Inactive Data Compression Scan started(sec): - Time since Last Successful Inactive Data Compression Scan ended(sec): 0 Average time for Cold Data Compression(sec): 0 Tuning Enabled: true Threshold: 14 Threshold Upper Limit: 21 Threshold Lower Limit: 14 Client Read history window: 14 Incompressible Data Percentage: 0% ::*> volume show -volume vol1* -fields available, filesystem-size, total, used, percent-used, performance-tier-inactive-user-data, performance-tier-inactive-user-data-percent, size, dedupe-space-saved, dedupe-space-saved-percent, dedupe-space-shared,logical-used, logical-used-percent,logical-used-by-afs, logical-available vserver volume size available filesystem-size total used percent-used dedupe-space-saved dedupe-space-saved-percent dedupe-space-shared logical-used logical-used-percent logical-available logical-used-by-afs performance-tier-inactive-user-data performance-tier-inactive-user-data-percent ------- ------ ---- --------- --------------- ------- ------ ------------ ------------------ -------------------------- ------------------- ------------ -------------------- ----------------- ------------------- ----------------------------------- ------------------------------------------- svm vol1 16GB 6.85GB 16GB 15.20GB 8.35GB 54% 808.8MB 9% 808.8MB 9.14GB 60% - 9.14GB - - svm2 vol1_dst 7.54GB 1.10GB 7.54GB 7.16GB 6.06GB 84% 4.98GB 45% 3.00GB 11.01GB 154% - 8.05GB 0B 0% svm2 vol1_dst2 9.84GB 1.45GB 9.84GB 9.35GB 7.90GB 84% 4.53GB 36% 4.10GB 12.40GB 133% - 9.06GB 0B 0% svm3 vol1_dst2_dst 9.88GB 1.55GB 9.88GB 9.88GB 8.32GB 84% 774.5MB 8% 8.24GB 9.04GB 92% - 9.04GB 0B 0% svm3 vol1_dst_dst 4.90GB 896.4MB 4.90GB 4.90GB 4.03GB 82% 5.11GB 56% 2.21GB 9.14GB 186% - 9.07GB 0B 0% 5 entries were displayed. ::*> volume show-footprint -volume vol1_dst2_dst Vserver : svm3 Volume : vol1_dst2_dst Feature Used Used% -------------------------------- ---------- ----- Volume Data Footprint 8.32GB 1% Footprint in Performance Tier 8.41GB 100% Footprint in FSxFabricpoolObjectStore 0B 0% Volume Guarantee 0B 0% Flexible Volume Metadata 61.31MB 0% Delayed Frees 86.66MB 0% File Operation Metadata 4KB 0% Total Footprint 8.46GB 1% Footprint Data Reduction 3.70GB 0% Auto Adaptive Compression 3.70GB 0% Effective Total Footprint 4.76GB 1% ::*> aggr show-efficiency -instance Name of the Aggregate: aggr1 Node where Aggregate Resides: FsxId0ab6f9b00824a187c-01 Logical Size Used by Volumes, Clones, Snapshot Copies in the Aggregate: 207.4GB Total Physical Used: 22.45GB Total Storage Efficiency Ratio: 9.24:1 Total Data Reduction Logical Used Without Snapshots: 43.97GB Total Data Reduction Physical Used Without Snapshots: 17.44GB Total Data Reduction Efficiency Ratio Without Snapshots: 2.52:1 Total Data Reduction Logical Used without snapshots and flexclones: 43.97GB Total Data Reduction Physical Used without snapshots and flexclones: 17.44GB Total Data Reduction Efficiency Ratio without snapshots and flexclones: 2.52:1 Total Logical Size Used by Volumes, Clones, Snapshot Copies in the FabricPool Performance Tier: 161.6GB Total Physical Used in FabricPool Performance Tier: 18.78GB Total FabricPool Performance Tier Storage Efficiency Ratio: 8.60:1 Total Data Reduction Logical Used without snapshots and flexclones in the FabricPool Performance Tier: 35.59GB Total Data Reduction Physical Used without snapshots and flexclones in the FabricPool Performance Tier: 13.80GB Total FabricPool Performance Tier Data Reduction Efficiency Ratio without snapshots and flexclones: 2.58:1 Logical Space Used for All Volumes: 43.97GB Physical Space Used for All Volumes: 27.80GB Space Saved by Volume Deduplication: 16.17GB Space Saved by Volume Deduplication and pattern detection: 16.17GB Volume Deduplication Savings ratio: 1.58:1 Space Saved by Volume Compression: 0B Volume Compression Savings ratio: 1.00:1 Space Saved by Inline Zero Pattern Detection: 0B Volume Data Reduction SE Ratio: 1.58:1 Logical Space Used by the Aggregate: 32.60GB Physical Space Used by the Aggregate: 22.45GB Space Saved by Aggregate Data Reduction: 10.15GB Aggregate Data Reduction SE Ratio: 1.45:1 Logical Size Used by Snapshot Copies: 163.4GB Physical Size Used by Snapshot Copies: 7.27GB Snapshot Volume Data Reduction Ratio: 22.48:1 Logical Size Used by FlexClone Volumes: 0B Physical Sized Used by FlexClone Volumes: 0B FlexClone Volume Data Reduction Ratio: 1.00:1 Snapshot And FlexClone Volume Data Reduction SE Ratio: 22.48:1 Number of Volumes Offline: 0 Number of SIS Disabled Volumes: 3 Number of SIS Change Log Disabled Volumes: 1 ::*> snapshot show -volume vol1* ---Blocks--- Vserver Volume Snapshot Size Total% Used% -------- -------- ------------------------------------- -------- ------ ----- svm vol1 test.2023-12-22_0533 160KB 0% 0% snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_053528 24.45MB 0% 0% snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_065900 312KB 0% 0% snapmirror.674509fa-a065-11ee-981e-bdd56ead09c8_2148779290.2023-12-22_075507 292KB 0% 0% snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779291.2023-12-25_050406 144KB 0% 0% svm2 vol1_dst snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_015623 1.91GB 25% 38% test.2023-12-22_0533 893.2MB 12% 22% snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_053528 481.9MB 6% 13% snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_065900 82.71MB 1% 3% vol1_dst2 test.2023-12-22_0533 608KB 0% 0% snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_053528 364KB 0% 0% snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_065900 420KB 0% 0% snapmirror.674509fa-a065-11ee-981e-bdd56ead09c8_2148779290.2023-12-22_075507 280KB 0% 0% snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779291.2023-12-25_050406 3.83GB 39% 46% svm3 vol1_dst2_dst test.2023-12-22_0533 536KB 0% 0% snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_053528 380KB 0% 0% snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_065900 380KB 0% 0% snapmirror.674509fa-a065-11ee-981e-bdd56ead09c8_2148779290.2023-12-22_075507 300KB 0% 0% snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779291.2023-12-25_050406 168KB 0% 0% vol1_dst_dst snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_065900 24.16MB 0% 1% snapmirror.674509fa-a065-11ee-981e-bdd56ead09c8_2148779290.2023-12-22_074801 41.57MB 1% 1% snapmirror.674509fa-a065-11ee-981e-bdd56ead09c8_2148779290.2023-12-22_075507 4.96MB 0% 0% 22 entries were displayed.
重複排除量もaggregateのデータ削減量も変化ありません。
また、vol1_dst2
にもvol1_dst2_dst
にもSnapshotは追加されていません。
そのため、Storage Efficiencyによる重複排除の最終的な転送先に反映するためにはvol1 -> vol1_dst -> vol1_dst_dst
とカスケードSnapMirrorを2回繰り返す必要があるようです。
svm と svm2 のボリューム間のSnapMirror update
svm と svm2 のボリューム間のSnapMirrorの差分転送を行います。
::*> snapmirror update -destination-path svm2:vol1_dst2 Operation is queued: snapmirror update of destination "svm2:vol1_dst2". ::*> snapmirror show Progress Source Destination Mirror Relationship Total Last Path Type Path State Status Progress Healthy Updated ----------- ---- ------------ ------- -------------- --------- ------- -------- svm:vol1 XDP svm2:vol1_dst Snapmirrored Idle - true - svm2:vol1_dst2 Snapmirrored Idle - true - svm3:vol1_dst_dst Broken-off Idle - true - svm2:vol1_dst2 XDP svm3:vol1_dst2_dst Snapmirrored Idle - true - 4 entries were displayed. ::*> snapmirror show -destination-path svm2:vol1_dst2 Source Path: svm:vol1 Source Cluster: - Source Vserver: svm Source Volume: vol1 Destination Path: svm2:vol1_dst2 Destination Cluster: - Destination Vserver: svm2 Destination Volume: vol1_dst2 Relationship Type: XDP Relationship Group Type: none Managing Vserver: svm2 SnapMirror Schedule: - SnapMirror Policy Type: async-mirror SnapMirror Policy: MirrorAllSnapshots Tries Limit: - Throttle (KB/sec): unlimited Consistency Group Item Mappings: - Current Transfer Throttle (KB/sec): - Mirror State: Snapmirrored Relationship Status: Idle File Restore File Count: - File Restore File List: - Transfer Snapshot: - Snapshot Progress: - Total Progress: - Network Compression Ratio: - Snapshot Checkpoint: - Newest Snapshot: snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779291.2023-12-25_060027 Newest Snapshot Timestamp: 12/25 06:00:27 Exported Snapshot: snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779291.2023-12-25_060027 Exported Snapshot Timestamp: 12/25 06:00:27 Healthy: true Relationship ID: 078ecbbc-a2e3-11ee-981e-bdd56ead09c8 Source Vserver UUID: 04ee1778-a058-11ee-981e-bdd56ead09c8 Destination Vserver UUID: 5af907bb-a065-11ee-981e-bdd56ead09c8 Current Operation ID: - Transfer Type: - Transfer Error: - Last Transfer Type: update Last Transfer Error: - Last Transfer Error Codes: - Last Transfer Size: 3.27KB Last Transfer Network Compression Ratio: 1:1 Last Transfer Duration: 0:0:2 Last Transfer From: svm:vol1 Last Transfer End Timestamp: 12/25 06:00:29 Unhealthy Reason: - Progress Last Updated: - Relationship Capability: 8.2 and above Lag Time: 0:0:22 Current Transfer Priority: - SMTape Operation: - Destination Volume Node Name: FsxId0ab6f9b00824a187c-01 Identity Preserve Vserver DR: - Number of Successful Updates: 2 Number of Failed Updates: 0 Number of Successful Resyncs: 0 Number of Failed Resyncs: 0 Number of Successful Breaks: 0 Number of Failed Breaks: 0 Total Transfer Bytes: 9057888528 Total Transfer Time in Seconds: 223 Source Volume MSIDs Preserved: - OpMask: ffffffffffffffff Is Auto Expand Enabled: - Percent Complete for Current Status: -
転送量は3.27KBです。
Storage Efficiency、ボリューム、aggregate、Snapshotの情報を確認します。
::*> volume efficiency show -volume vol1* -fields state, policy, storage-efficiency-mode, inline-compression, inline-dedupe, compression, data-compaction, auto-adaptive-compression-savings, auto-adaptive-compression-existing-volume, using-auto-adaptive-compression vserver volume state policy compression inline-compression storage-efficiency-mode inline-dedupe data-compaction auto-adaptive-compression-savings using-auto-adaptive-compression auto-adaptive-compression-existing-volume ------- ------ ------- ------ ----------- ------------------ ----------------------- ------------- --------------- --------------------------------- ------------------------------- ----------------------------------------- svm vol1 Enabled auto false true efficient false true true true false svm2 vol1_dst Enabled - false true efficient false true true true false svm2 vol1_dst2 Enabled - false true efficient false true true true false svm3 vol1_dst2_dst Disabled - false true efficient false true true true false svm3 vol1_dst_dst Enabled auto false true efficient true true true true false 5 entries were displayed. ::*> volume efficiency inactive-data-compression show -volume vol1_dst2_dst -instance Volume: vol1_dst2_dst Vserver: svm3 Is Enabled: true Scan Mode: - Progress: IDLE Status: SUCCESS Compression Algorithm: lzopro Failure Reason: - Total Blocks: - Total blocks Processed: - Percentage: - Phase1 L1s Processed: - Phase1 Lns Skipped: - Phase2 Total Blocks: - Phase2 Blocks Processed: - Number of Cold Blocks Encountered: 0 Number of Repacked Blocks: 0 Number of Compression Done Blocks: 0 Number of Vol-Overwrites: 0 Time since Last Inactive Data Compression Scan started(sec): 0 Time since Last Inactive Data Compression Scan ended(sec): 878 Time since Last Successful Inactive Data Compression Scan started(sec): - Time since Last Successful Inactive Data Compression Scan ended(sec): 0 Average time for Cold Data Compression(sec): 0 Tuning Enabled: true Threshold: 14 Threshold Upper Limit: 21 Threshold Lower Limit: 14 Client Read history window: 14 Incompressible Data Percentage: 0% ::*> volume show -volume vol1* -fields available, filesystem-size, total, used, percent-used, performance-tier-inactive-user-data, performance-tier-inactive-user-data-percent, size, dedupe-space-saved, dedupe-space-saved-percent, dedupe-space-shared,logical-used, logical-used-percent,logical-used-by-afs, logical-available vserver volume size available filesystem-size total used percent-used dedupe-space-saved dedupe-space-saved-percent dedupe-space-shared logical-used logical-used-percent logical-available logical-used-by-afs performance-tier-inactive-user-data performance-tier-inactive-user-data-percent ------- ------ ---- --------- --------------- ------- ------ ------------ ------------------ -------------------------- ------------------- ------------ -------------------- ----------------- ------------------- ----------------------------------- ------------------------------------------- svm vol1 16GB 6.85GB 16GB 15.20GB 8.35GB 54% 808.8MB 9% 808.8MB 9.14GB 60% - 9.14GB - - svm2 vol1_dst 7.54GB 1.10GB 7.54GB 7.16GB 6.06GB 84% 4.98GB 45% 3.00GB 11.01GB 154% - 8.05GB 0B 0% svm2 vol1_dst2 9.84GB 1.45GB 9.84GB 9.35GB 7.90GB 84% 4.53GB 36% 4.10GB 12.40GB 133% - 9.06GB 0B 0% svm3 vol1_dst2_dst 9.88GB 1.55GB 9.88GB 9.88GB 8.32GB 84% 774.5MB 8% 8.24GB 9.04GB 92% - 9.04GB 0B 0% svm3 vol1_dst_dst 4.90GB 896.4MB 4.90GB 4.90GB 4.03GB 82% 5.11GB 56% 2.21GB 9.14GB 186% - 9.07GB 0B 0% 5 entries were displayed. ::*> volume show-footprint -volume vol1_dst2_dst Vserver : svm3 Volume : vol1_dst2_dst Feature Used Used% -------------------------------- ---------- ----- Volume Data Footprint 8.32GB 1% Footprint in Performance Tier 8.41GB 100% Footprint in FSxFabricpoolObjectStore 0B 0% Volume Guarantee 0B 0% Flexible Volume Metadata 61.31MB 0% Delayed Frees 86.66MB 0% File Operation Metadata 4KB 0% Total Footprint 8.46GB 1% Footprint Data Reduction 3.70GB 0% Auto Adaptive Compression 3.70GB 0% Effective Total Footprint 4.76GB 1% ::*> aggr show-efficiency -instance Name of the Aggregate: aggr1 Node where Aggregate Resides: FsxId0ab6f9b00824a187c-01 Logical Size Used by Volumes, Clones, Snapshot Copies in the Aggregate: 225.6GB Total Physical Used: 22.45GB Total Storage Efficiency Ratio: 10.05:1 Total Data Reduction Logical Used Without Snapshots: 43.96GB Total Data Reduction Physical Used Without Snapshots: 17.45GB Total Data Reduction Efficiency Ratio Without Snapshots: 2.52:1 Total Data Reduction Logical Used without snapshots and flexclones: 43.96GB Total Data Reduction Physical Used without snapshots and flexclones: 17.45GB Total Data Reduction Efficiency Ratio without snapshots and flexclones: 2.52:1 Total Logical Size Used by Volumes, Clones, Snapshot Copies in the FabricPool Performance Tier: 170.9GB Total Physical Used in FabricPool Performance Tier: 18.79GB Total FabricPool Performance Tier Storage Efficiency Ratio: 9.10:1 Total Data Reduction Logical Used without snapshots and flexclones in the FabricPool Performance Tier: 35.59GB Total Data Reduction Physical Used without snapshots and flexclones in the FabricPool Performance Tier: 13.80GB Total FabricPool Performance Tier Data Reduction Efficiency Ratio without snapshots and flexclones: 2.58:1 Logical Space Used for All Volumes: 43.96GB Physical Space Used for All Volumes: 27.80GB Space Saved by Volume Deduplication: 16.17GB Space Saved by Volume Deduplication and pattern detection: 16.17GB Volume Deduplication Savings ratio: 1.58:1 Space Saved by Volume Compression: 0B Volume Compression Savings ratio: 1.00:1 Space Saved by Inline Zero Pattern Detection: 0B Volume Data Reduction SE Ratio: 1.58:1 Logical Space Used by the Aggregate: 32.60GB Physical Space Used by the Aggregate: 22.45GB Space Saved by Aggregate Data Reduction: 10.15GB Aggregate Data Reduction SE Ratio: 1.45:1 Logical Size Used by Snapshot Copies: 181.6GB Physical Size Used by Snapshot Copies: 7.27GB Snapshot Volume Data Reduction Ratio: 24.98:1 Logical Size Used by FlexClone Volumes: 0B Physical Sized Used by FlexClone Volumes: 0B FlexClone Volume Data Reduction Ratio: 1.00:1 Snapshot And FlexClone Volume Data Reduction SE Ratio: 24.98:1 Number of Volumes Offline: 0 Number of SIS Disabled Volumes: 3 Number of SIS Change Log Disabled Volumes: 1 ::*> snapshot show -volume vol1* ---Blocks--- Vserver Volume Snapshot Size Total% Used% -------- -------- ------------------------------------- -------- ------ ----- svm vol1 test.2023-12-22_0533 160KB 0% 0% snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_053528 24.45MB 0% 0% snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_065900 312KB 0% 0% snapmirror.674509fa-a065-11ee-981e-bdd56ead09c8_2148779290.2023-12-22_075507 292KB 0% 0% snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779291.2023-12-25_050406 148KB 0% 0% snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779291.2023-12-25_060027 136KB 0% 0% svm2 vol1_dst snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_015623 1.91GB 25% 38% test.2023-12-22_0533 893.2MB 12% 22% snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_053528 481.9MB 6% 13% snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_065900 82.71MB 1% 3% vol1_dst2 test.2023-12-22_0533 608KB 0% 0% snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_053528 364KB 0% 0% snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_065900 420KB 0% 0% snapmirror.674509fa-a065-11ee-981e-bdd56ead09c8_2148779290.2023-12-22_075507 280KB 0% 0% snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779291.2023-12-25_050406 3.83GB 39% 46% snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779291.2023-12-25_060027 140KB 0% 0% svm3 vol1_dst2_dst test.2023-12-22_0533 536KB 0% 0% snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_053528 380KB 0% 0% snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_065900 380KB 0% 0% snapmirror.674509fa-a065-11ee-981e-bdd56ead09c8_2148779290.2023-12-22_075507 300KB 0% 0% snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779291.2023-12-25_050406 168KB 0% 0% vol1_dst_dst snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_065900 24.16MB 0% 1% snapmirror.674509fa-a065-11ee-981e-bdd56ead09c8_2148779290.2023-12-22_074801 41.57MB 1% 1% snapmirror.674509fa-a065-11ee-981e-bdd56ead09c8_2148779290.2023-12-22_075507 4.96MB 0% 0% 24 entries were displayed.
vol1
とvol1_dst2
にSnapshotが追加されていますね。
svm2 と svm3 のボリューム間のSnapMirror update
svm2 と svm3 のボリューム間のSnapMirrorの差分転送を行います。
::*> snapmirror update -destination-path svm3:vol1_dst2_dst Operation is queued: snapmirror update of destination "svm3:vol1_dst2_dst". ::*> snapmirror show Progress Source Destination Mirror Relationship Total Last Path Type Path State Status Progress Healthy Updated ----------- ---- ------------ ------- -------------- --------- ------- -------- svm:vol1 XDP svm2:vol1_dst Snapmirrored Idle - true - svm2:vol1_dst2 Snapmirrored Idle - true - svm3:vol1_dst_dst Broken-off Idle - true - svm2:vol1_dst2 XDP svm3:vol1_dst2_dst Snapmirrored Finalizing 70.43MB true 12/25 06:01:34 4 entries were displayed. ::*> snapmirror show Progress Source Destination Mirror Relationship Total Last Path Type Path State Status Progress Healthy Updated ----------- ---- ------------ ------- -------------- --------- ------- -------- svm:vol1 XDP svm2:vol1_dst Snapmirrored Idle - true - svm2:vol1_dst2 Snapmirrored Idle - true - svm3:vol1_dst_dst Broken-off Idle - true - svm2:vol1_dst2 XDP svm3:vol1_dst2_dst Snapmirrored Idle - true - 4 entries were displayed. ::*> snapmirror show -destination-path svm3:vol1_dst2_dst Source Path: svm2:vol1_dst2 Source Cluster: - Source Vserver: svm2 Source Volume: vol1_dst2 Destination Path: svm3:vol1_dst2_dst Destination Cluster: - Destination Vserver: svm3 Destination Volume: vol1_dst2_dst Relationship Type: XDP Relationship Group Type: none Managing Vserver: svm3 SnapMirror Schedule: - SnapMirror Policy Type: async-mirror SnapMirror Policy: MirrorAllSnapshots Tries Limit: - Throttle (KB/sec): unlimited Consistency Group Item Mappings: - Current Transfer Throttle (KB/sec): - Mirror State: Snapmirrored Relationship Status: Idle File Restore File Count: - File Restore File List: - Transfer Snapshot: - Snapshot Progress: - Total Progress: - Network Compression Ratio: - Snapshot Checkpoint: - Newest Snapshot: snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779291.2023-12-25_060027 Newest Snapshot Timestamp: 12/25 06:00:27 Exported Snapshot: snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779291.2023-12-25_060027 Exported Snapshot Timestamp: 12/25 06:00:27 Healthy: true Relationship ID: cb1dc5b5-a2e8-11ee-981e-bdd56ead09c8 Source Vserver UUID: 5af907bb-a065-11ee-981e-bdd56ead09c8 Destination Vserver UUID: 674509fa-a065-11ee-981e-bdd56ead09c8 Current Operation ID: - Transfer Type: - Transfer Error: - Last Transfer Type: update Last Transfer Error: - Last Transfer Error Codes: - Last Transfer Size: 70.43MB Last Transfer Network Compression Ratio: 1:1 Last Transfer Duration: 0:0:26 Last Transfer From: svm2:vol1_dst2 Last Transfer End Timestamp: 12/25 06:01:55 Unhealthy Reason: - Progress Last Updated: - Relationship Capability: 8.2 and above Lag Time: 0:1:39 Current Transfer Priority: - SMTape Operation: - Destination Volume Node Name: FsxId0ab6f9b00824a187c-01 Identity Preserve Vserver DR: - Number of Successful Updates: 3 Number of Failed Updates: 0 Number of Successful Resyncs: 0 Number of Failed Resyncs: 0 Number of Successful Breaks: 0 Number of Failed Breaks: 0 Total Transfer Bytes: 5964877057 Total Transfer Time in Seconds: 102 Source Volume MSIDs Preserved: - OpMask: ffffffffffffffff Is Auto Expand Enabled: - Percent Complete for Current Status: -
転送量は70.43MBでした。
Storage Efficiency、ボリューム、aggregate、Snapshotの情報を確認します。
::*> volume efficiency show -volume vol1* -fields state, policy, storage-efficiency-mode, inline-compression, inline-dedupe, compression, data-compaction, auto-adaptive-compression-savings, auto-adaptive-compression-existing-volume, using-auto-adaptive-compression vserver volume state policy compression inline-compression storage-efficiency-mode inline-dedupe data-compaction auto-adaptive-compression-savings using-auto-adaptive-compression auto-adaptive-compression-existing-volume ------- ------ ------- ------ ----------- ------------------ ----------------------- ------------- --------------- --------------------------------- ------------------------------- ----------------------------------------- svm vol1 Enabled auto false true efficient false true true true false svm2 vol1_dst Enabled - false true efficient false true true true false svm2 vol1_dst2 Enabled - false true efficient false true true true false svm3 vol1_dst2_dst Disabled - false true efficient false true true true false svm3 vol1_dst_dst Enabled auto false true efficient true true true true false 5 entries were displayed. ::*> volume efficiency inactive-data-compression show -volume vol1_dst2_dst -instance Volume: vol1_dst2_dst Vserver: svm3 Is Enabled: true Scan Mode: - Progress: IDLE Status: SUCCESS Compression Algorithm: lzopro Failure Reason: - Total Blocks: - Total blocks Processed: - Percentage: - Phase1 L1s Processed: - Phase1 Lns Skipped: - Phase2 Total Blocks: - Phase2 Blocks Processed: - Number of Cold Blocks Encountered: 0 Number of Repacked Blocks: 0 Number of Compression Done Blocks: 0 Number of Vol-Overwrites: 0 Time since Last Inactive Data Compression Scan started(sec): 0 Time since Last Inactive Data Compression Scan ended(sec): 953 Time since Last Successful Inactive Data Compression Scan started(sec): - Time since Last Successful Inactive Data Compression Scan ended(sec): 0 Average time for Cold Data Compression(sec): 0 Tuning Enabled: true Threshold: 14 Threshold Upper Limit: 21 Threshold Lower Limit: 14 Client Read history window: 14 Incompressible Data Percentage: 0% ::*> volume show -volume vol1* -fields available, filesystem-size, total, used, percent-used, performance-tier-inactive-user-data, performance-tier-inactive-user-data-percent, size, dedupe-space-saved, dedupe-space-saved-percent, dedupe-space-shared,logical-used, logical-used-percent,logical-used-by-afs, logical-available vserver volume size available filesystem-size total used percent-used dedupe-space-saved dedupe-space-saved-percent dedupe-space-shared logical-used logical-used-percent logical-available logical-used-by-afs performance-tier-inactive-user-data performance-tier-inactive-user-data-percent ------- ------ ---- --------- --------------- ------- ------ ------------ ------------------ -------------------------- ------------------- ------------ -------------------- ----------------- ------------------- ----------------------------------- ------------------------------------------- svm vol1 16GB 6.85GB 16GB 15.20GB 8.35GB 54% 808.8MB 9% 808.8MB 9.14GB 60% - 9.14GB - - svm2 vol1_dst 7.54GB 1.10GB 7.54GB 7.16GB 6.06GB 84% 4.98GB 45% 3.00GB 11.01GB 154% - 8.05GB 0B 0% svm2 vol1_dst2 9.84GB 1.45GB 9.84GB 9.35GB 7.90GB 84% 4.53GB 36% 4.10GB 12.40GB 133% - 9.06GB 0B 0% svm3 vol1_dst2_dst 10.00GB 1.63GB 10.00GB 10.00GB 8.37GB 83% 4.53GB 35% 4.41GB 12.87GB 129% - 9.04GB 0B 0% svm3 vol1_dst_dst 4.90GB 896.4MB 4.90GB 4.90GB 4.03GB 82% 5.11GB 56% 2.21GB 9.14GB 186% - 9.07GB 0B 0% 5 entries were displayed. ::*> volume show-footprint -volume vol1_dst2_dst Vserver : svm3 Volume : vol1_dst2_dst Feature Used Used% -------------------------------- ---------- ----- Volume Data Footprint 8.37GB 1% Footprint in Performance Tier 8.47GB 100% Footprint in FSxFabricpoolObjectStore 0B 0% Volume Guarantee 0B 0% Flexible Volume Metadata 61.92MB 0% Delayed Frees 100.2MB 0% File Operation Metadata 4KB 0% Total Footprint 8.53GB 1% Footprint Data Reduction 3.73GB 0% Auto Adaptive Compression 3.73GB 0% Effective Total Footprint 4.80GB 1% ::*> aggr show-efficiency -instance Name of the Aggregate: aggr1 Node where Aggregate Resides: FsxId0ab6f9b00824a187c-01 Logical Size Used by Volumes, Clones, Snapshot Copies in the Aggregate: 234.6GB Total Physical Used: 22.51GB Total Storage Efficiency Ratio: 10.42:1 Total Data Reduction Logical Used Without Snapshots: 43.96GB Total Data Reduction Physical Used Without Snapshots: 14.86GB Total Data Reduction Efficiency Ratio Without Snapshots: 2.96:1 Total Data Reduction Logical Used without snapshots and flexclones: 43.96GB Total Data Reduction Physical Used without snapshots and flexclones: 14.86GB Total Data Reduction Efficiency Ratio without snapshots and flexclones: 2.96:1 Total Logical Size Used by Volumes, Clones, Snapshot Copies in the FabricPool Performance Tier: 180.0GB Total Physical Used in FabricPool Performance Tier: 18.85GB Total FabricPool Performance Tier Storage Efficiency Ratio: 9.55:1 Total Data Reduction Logical Used without snapshots and flexclones in the FabricPool Performance Tier: 35.59GB Total Data Reduction Physical Used without snapshots and flexclones in the FabricPool Performance Tier: 11.22GB Total FabricPool Performance Tier Data Reduction Efficiency Ratio without snapshots and flexclones: 3.17:1 Logical Space Used for All Volumes: 43.96GB Physical Space Used for All Volumes: 24.02GB Space Saved by Volume Deduplication: 19.95GB Space Saved by Volume Deduplication and pattern detection: 19.95GB Volume Deduplication Savings ratio: 1.83:1 Space Saved by Volume Compression: 0B Volume Compression Savings ratio: 1.00:1 Space Saved by Inline Zero Pattern Detection: 0B Volume Data Reduction SE Ratio: 1.83:1 Logical Space Used by the Aggregate: 32.66GB Physical Space Used by the Aggregate: 22.51GB Space Saved by Aggregate Data Reduction: 10.15GB Aggregate Data Reduction SE Ratio: 1.45:1 Logical Size Used by Snapshot Copies: 190.6GB Physical Size Used by Snapshot Copies: 11.09GB Snapshot Volume Data Reduction Ratio: 17.18:1 Logical Size Used by FlexClone Volumes: 0B Physical Sized Used by FlexClone Volumes: 0B FlexClone Volume Data Reduction Ratio: 1.00:1 Snapshot And FlexClone Volume Data Reduction SE Ratio: 17.18:1 Number of Volumes Offline: 0 Number of SIS Disabled Volumes: 3 Number of SIS Change Log Disabled Volumes: 1 ::*> snapshot show -volume vol1* ---Blocks--- Vserver Volume Snapshot Size Total% Used% -------- -------- ------------------------------------- -------- ------ ----- svm vol1 test.2023-12-22_0533 160KB 0% 0% snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_053528 24.45MB 0% 0% snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_065900 312KB 0% 0% snapmirror.674509fa-a065-11ee-981e-bdd56ead09c8_2148779290.2023-12-22_075507 292KB 0% 0% snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779291.2023-12-25_050406 148KB 0% 0% snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779291.2023-12-25_060027 136KB 0% 0% svm2 vol1_dst snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_015623 1.91GB 25% 38% test.2023-12-22_0533 893.2MB 12% 22% snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_053528 481.9MB 6% 13% snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_065900 82.71MB 1% 3% vol1_dst2 test.2023-12-22_0533 608KB 0% 0% snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_053528 364KB 0% 0% snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_065900 420KB 0% 0% snapmirror.674509fa-a065-11ee-981e-bdd56ead09c8_2148779290.2023-12-22_075507 280KB 0% 0% snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779291.2023-12-25_050406 3.83GB 39% 46% snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779291.2023-12-25_060027 140KB 0% 0% svm3 vol1_dst2_dst test.2023-12-22_0533 536KB 0% 0% snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_053528 380KB 0% 0% snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_065900 380KB 0% 0% snapmirror.674509fa-a065-11ee-981e-bdd56ead09c8_2148779290.2023-12-22_075507 300KB 0% 0% snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779291.2023-12-25_050406 3.83GB 38% 46% snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779291.2023-12-25_060027 148KB 0% 0% vol1_dst_dst snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_065900 24.16MB 0% 1% snapmirror.674509fa-a065-11ee-981e-bdd56ead09c8_2148779290.2023-12-22_074801 41.57MB 1% 1% snapmirror.674509fa-a065-11ee-981e-bdd56ead09c8_2148779290.2023-12-22_075507 4.96MB 0% 0% 25 entries were displayed.
vol1_dst2_dst
で重複排除が効いていますね。
そのため、Storage Efficiencyによる重複排除の最終的な転送先に反映するためにはvol1 -> vol1_dst -> vol1_dst_dst
とカスケードSnapMirrorを2回繰り返す必要があることが分かりました。
vol1_dst2_dst のSnapshotを削除する
vol1_dst2_dst
のSnapshotを削除していってボリュームやaggregateの使用量の変化を確認します。
SnapMirrorをカットオーバーします。
::*> snapmirror quiesce -destination-path svm3:vol1_dst2_dst Operation succeeded: snapmirror quiesce for destination "svm3:vol1_dst2_dst". ::*> snapmirror break -destination-path svm3:vol1_dst2_dst Operation succeeded: snapmirror break for destination "svm3:vol1_dst2_dst". ::*> snapmirror show Progress Source Destination Mirror Relationship Total Last Path Type Path State Status Progress Healthy Updated ----------- ---- ------------ ------- -------------- --------- ------- -------- svm:vol1 XDP svm2:vol1_dst Snapmirrored Idle - true - svm2:vol1_dst2 Snapmirrored Idle - true - svm3:vol1_dst_dst Broken-off Idle - true - svm2:vol1_dst2 XDP svm3:vol1_dst2_dst Broken-off Idle - true - 4 entries were displayed. ::*> volume efficiency show -volume vol1* -fields state, policy, storage-efficiency-mode, inline-compression, inline-dedupe, compression, data-compaction, auto-adaptive-compression-savings, auto-adaptive-compression-existing-volume,using-auto-adaptive-compression vserver volume state policy compression inline-compression storage-efficiency-mode inline-dedupe data-compaction auto-adaptive-compression-savings using-auto-adaptive-compression auto-adaptive-compression-existing-volume ------- ------ ------- ------ ----------- ------------------ ----------------------- ------------- --------------- --------------------------------- ------------------------------- ----------------------------------------- svm vol1 Enabled auto false true efficient false true true true false svm2 vol1_dst Enabled - false true efficient false true true true false svm2 vol1_dst2 Enabled - false true efficient false true true true false svm3 vol1_dst2_dst Enabled auto false true efficient true true true true false svm3 vol1_dst_dst Enabled auto false true efficient true true true true false 5 entries were displayed.
`vol1_dst2_dstをマウントします。
::*> volume mount -vserver svm3 -volume vol1_dst2_dst -junction-path /vol1_dst2_dst Queued private job: 52
NFSクライアントからvol1_dst2_dst
をマウントします。
$ sudo mkdir -p /mnt/fsxn/vol1_dst2_dst 2$ sudo mount -t nfs svm-00c880c6c6eb922ed.fs-0ab6f9b00824a187c.fsx.us-east-1.amazonaws.com:/vol1_dst2_dst /mnt/fsxn/vol1_dst2_dst $ df -hT -t nfs4 Filesystem Type Size Used Avail Use% Mounted on svm-0058ae83d258ab2e3.fs-0ab6f9b00824a187c.fsx.us-east-1.amazonaws.com:/vol1 nfs4 16G 8.4G 6.9G 55% /mnt/fsxn/vol1 svm-00c880c6c6eb922ed.fs-0ab6f9b00824a187c.fsx.us-east-1.amazonaws.com:/vol1_dst_dst nfs4 5.0G 4.1G 897M 83% /mnt/fsxn/vol1_dst_dst svm-00c880c6c6eb922ed.fs-0ab6f9b00824a187c.fsx.us-east-1.amazonaws.com:/vol1_dst2_dst nfs4 10G 8.4G 1.7G 84% /mnt/fsxn/vol1_dst2_dst $ ls -l /mnt/fsxn/vol1_dst2_dst/.snapshot total 24 drwxr-xr-x. 2 root root 4096 Dec 22 05:28 snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_053528 drwxr-xr-x. 2 root root 4096 Dec 22 06:55 snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_065900 drwxr-xr-x. 2 root root 4096 Dec 22 07:53 snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779291.2023-12-25_050406 drwxr-xr-x. 2 root root 4096 Dec 22 07:53 snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779291.2023-12-25_060027 drwxr-xr-x. 2 root root 4096 Dec 22 07:53 snapmirror.674509fa-a065-11ee-981e-bdd56ead09c8_2148779290.2023-12-22_075507 drwxr-xr-x. 2 root root 4096 Dec 22 05:28 test.2023-12-22_0533
Snapshotを削除します。
::*> snapshot delete -vserver svm3 -volume vol1_dst2_dst -snapshot test.2023-12-22_0533 Warning: Deleting a Snapshot copy permanently removes data that is stored only in that Snapshot copy. Are you sure you want to delete Snapshot copy "test.2023-12-22_0533" for volume "vol1_dst2_dst" in Vserver "svm3" ? {y|n}: y ::*> ::*> snapshot show -volume vol1* ---Blocks--- Vserver Volume Snapshot Size Total% Used% -------- -------- ------------------------------------- -------- ------ ----- svm vol1 test.2023-12-22_0533 160KB 0% 0% snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_053528 24.45MB 0% 0% snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_065900 312KB 0% 0% snapmirror.674509fa-a065-11ee-981e-bdd56ead09c8_2148779290.2023-12-22_075507 292KB 0% 0% snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779291.2023-12-25_050406 148KB 0% 0% snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779291.2023-12-25_060027 136KB 0% 0% svm2 vol1_dst snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_015623 1.91GB 25% 38% test.2023-12-22_0533 893.2MB 12% 22% snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_053528 481.9MB 6% 13% snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_065900 82.71MB 1% 3% vol1_dst2 test.2023-12-22_0533 608KB 0% 0% snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_053528 364KB 0% 0% snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_065900 420KB 0% 0% snapmirror.674509fa-a065-11ee-981e-bdd56ead09c8_2148779290.2023-12-22_075507 280KB 0% 0% snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779291.2023-12-25_050406 3.83GB 39% 46% snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779291.2023-12-25_060027 140KB 0% 0% svm3 vol1_dst2_dst snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_053528 380KB 0% 0% snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_065900 380KB 0% 0% snapmirror.674509fa-a065-11ee-981e-bdd56ead09c8_2148779290.2023-12-22_075507 300KB 0% 0% snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779291.2023-12-25_050406 3.83GB 38% 46% snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779291.2023-12-25_060027 93.90MB 1% 2% vol1_dst_dst snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_065900 24.16MB 0% 1% snapmirror.674509fa-a065-11ee-981e-bdd56ead09c8_2148779290.2023-12-22_074801 41.57MB 1% 1% snapmirror.674509fa-a065-11ee-981e-bdd56ead09c8_2148779290.2023-12-22_075507 4.96MB 0% 0% 24 entries were displayed.
Storage Efficiency、ボリューム、aggregate、Snapshotの情報を確認します。
::*> volume show -volume vol1* -fields available, filesystem-size, total, used, percent-used, performance-tier-inactive-user-data, performance-tier-inactive-user-data-percent, size, dedupe-space-saved, dedupe-space-saved-percent, dedupe-space-shared,logical-used, logical-used-percent,logical-used-by-afs, logical-available vserver volume size available filesystem-size total used percent-used dedupe-space-saved dedupe-space-saved-percent dedupe-space-shared logical-used logical-used-percent logical-available logical-used-by-afs performance-tier-inactive-user-data performance-tier-inactive-user-data-percent ------- ------ ---- --------- --------------- ------- ------ ------------ ------------------ -------------------------- ------------------- ------------ -------------------- ----------------- ------------------- ----------------------------------- ------------------------------------------- svm vol1 16GB 6.85GB 16GB 15.20GB 8.35GB 54% 808.8MB 9% 808.8MB 9.14GB 60% - 9.14GB - - svm2 vol1_dst 7.54GB 1.10GB 7.54GB 7.16GB 6.06GB 84% 4.98GB 45% 3.00GB 11.01GB 154% - 8.05GB 0B 0% svm2 vol1_dst2 9.84GB 1.45GB 9.84GB 9.35GB 7.90GB 84% 4.53GB 36% 4.10GB 12.40GB 133% - 9.06GB 0B 0% svm3 vol1_dst2_dst 10.00GB 1.62GB 10.00GB 10.00GB 8.37GB 83% 4.68GB 36% 2.15GB 13.06GB 131% - 9.14GB 0B 0% svm3 vol1_dst_dst 4.90GB 896.4MB 4.90GB 4.90GB 4.03GB 82% 5.11GB 56% 2.21GB 9.14GB 186% - 9.07GB 0B 0% 5 entries were displayed. ::*> volume show-footprint -volume vol1_dst2_dst Vserver : svm3 Volume : vol1_dst2_dst Feature Used Used% -------------------------------- ---------- ----- Volume Data Footprint 8.37GB 1% Footprint in Performance Tier 8.46GB 100% Footprint in FSxFabricpoolObjectStore 0B 0% Volume Guarantee 0B 0% Flexible Volume Metadata 61.92MB 0% Delayed Frees 91.20MB 0% File Operation Metadata 4KB 0% Total Footprint 8.52GB 1% Footprint Data Reduction 3.73GB 0% Auto Adaptive Compression 3.73GB 0% Effective Total Footprint 4.80GB 1% ::*> aggr show-efficiency -instance Name of the Aggregate: aggr1 Node where Aggregate Resides: FsxId0ab6f9b00824a187c-01 Logical Size Used by Volumes, Clones, Snapshot Copies in the Aggregate: 228.6GB Total Physical Used: 22.52GB Total Storage Efficiency Ratio: 10.15:1 Total Data Reduction Logical Used Without Snapshots: 44.03GB Total Data Reduction Physical Used Without Snapshots: 14.81GB Total Data Reduction Efficiency Ratio Without Snapshots: 2.97:1 Total Data Reduction Logical Used without snapshots and flexclones: 44.03GB Total Data Reduction Physical Used without snapshots and flexclones: 14.81GB Total Data Reduction Efficiency Ratio without snapshots and flexclones: 2.97:1 Total Logical Size Used by Volumes, Clones, Snapshot Copies in the FabricPool Performance Tier: 174.0GB Total Physical Used in FabricPool Performance Tier: 18.86GB Total FabricPool Performance Tier Storage Efficiency Ratio: 9.23:1 Total Data Reduction Logical Used without snapshots and flexclones in the FabricPool Performance Tier: 35.65GB Total Data Reduction Physical Used without snapshots and flexclones in the FabricPool Performance Tier: 11.16GB Total FabricPool Performance Tier Data Reduction Efficiency Ratio without snapshots and flexclones: 3.19:1 Logical Space Used for All Volumes: 44.03GB Physical Space Used for All Volumes: 23.93GB Space Saved by Volume Deduplication: 20.09GB Space Saved by Volume Deduplication and pattern detection: 20.09GB Volume Deduplication Savings ratio: 1.84:1 Space Saved by Volume Compression: 0B Volume Compression Savings ratio: 1.00:1 Space Saved by Inline Zero Pattern Detection: 0B Volume Data Reduction SE Ratio: 1.84:1 Logical Space Used by the Aggregate: 32.67GB Physical Space Used by the Aggregate: 22.52GB Space Saved by Aggregate Data Reduction: 10.15GB Aggregate Data Reduction SE Ratio: 1.45:1 Logical Size Used by Snapshot Copies: 184.6GB Physical Size Used by Snapshot Copies: 11.19GB Snapshot Volume Data Reduction Ratio: 16.50:1 Logical Size Used by FlexClone Volumes: 0B Physical Sized Used by FlexClone Volumes: 0B FlexClone Volume Data Reduction Ratio: 1.00:1 Snapshot And FlexClone Volume Data Reduction SE Ratio: 16.50:1 Number of Volumes Offline: 0 Number of SIS Disabled Volumes: 3 Number of SIS Change Log Disabled Volumes: 0
536KBのSnapshotを削除したからか特に変化はありません。
NFSクライアントからも使用量を確認します。
$ df -hT -t nfs4 Filesystem Type Size Used Avail Use% Mounted on svm-0058ae83d258ab2e3.fs-0ab6f9b00824a187c.fsx.us-east-1.amazonaws.com:/vol1 nfs4 16G 8.4G 6.9G 55% /mnt/fsxn/vol1 svm-00c880c6c6eb922ed.fs-0ab6f9b00824a187c.fsx.us-east-1.amazonaws.com:/vol1_dst_dst nfs4 5.0G 4.1G 897M 83% /mnt/fsxn/vol1_dst_dst svm-00c880c6c6eb922ed.fs-0ab6f9b00824a187c.fsx.us-east-1.amazonaws.com:/vol1_dst2_dst nfs4 10G 8.4G 1.7G 84% /mnt/fsxn/vol1_dst2_dst
こちらも特に変化ありません。
300KBほどのSnapshotをまとめて削除してみます。
::*> snapshot delete -vserver svm3 -volume vol1_dst2_dst -snapshot snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_053528, snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_065900, snapmirror.674509fa-a065-11ee-981e-bdd56ead09c8_2148779290.2023-12-22_075507 Warning: Deleting a Snapshot copy permanently removes data that is stored only in that Snapshot copy. Are you sure you want to delete Snapshot copy "snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_053528" for volume "vol1_dst2_dst" in Vserver "svm3" ? {y|n}: y Warning: Deleting a Snapshot copy permanently removes data that is stored only in that Snapshot copy. Are you sure you want to delete Snapshot copy "snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_065900" for volume "vol1_dst2_dst" in Vserver "svm3" ? {y|n}: y Warning: Deleting a Snapshot copy permanently removes data that is stored only in that Snapshot copy. Are you sure you want to delete Snapshot copy "snapmirror.674509fa-a065-11ee-981e-bdd56ead09c8_2148779290.2023-12-22_075507" for volume "vol1_dst2_dst" in Vserver "svm3" ? {y|n}: y 3 entries were acted on. ::*> snapshot show -volume vol1* ---Blocks--- Vserver Volume Snapshot Size Total% Used% -------- -------- ------------------------------------- -------- ------ ----- svm vol1 test.2023-12-22_0533 160KB 0% 0% snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_053528 24.45MB 0% 0% snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_065900 312KB 0% 0% snapmirror.674509fa-a065-11ee-981e-bdd56ead09c8_2148779290.2023-12-22_075507 292KB 0% 0% snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779291.2023-12-25_050406 148KB 0% 0% snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779291.2023-12-25_060027 136KB 0% 0% svm2 vol1_dst snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_015623 1.91GB 25% 38% test.2023-12-22_0533 893.2MB 12% 22% snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_053528 481.9MB 6% 13% snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_065900 82.71MB 1% 3% vol1_dst2 test.2023-12-22_0533 608KB 0% 0% snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_053528 364KB 0% 0% snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_065900 420KB 0% 0% snapmirror.674509fa-a065-11ee-981e-bdd56ead09c8_2148779290.2023-12-22_075507 280KB 0% 0% snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779291.2023-12-25_050406 3.83GB 39% 46% snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779291.2023-12-25_060027 140KB 0% 0% svm3 vol1_dst2_dst snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779291.2023-12-25_050406 3.83GB 38% 46% snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779291.2023-12-25_060027 93.90MB 1% 2% vol1_dst_dst snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_065900 24.16MB 0% 1% snapmirror.674509fa-a065-11ee-981e-bdd56ead09c8_2148779290.2023-12-22_074801 41.57MB 1% 1% snapmirror.674509fa-a065-11ee-981e-bdd56ead09c8_2148779290.2023-12-22_075507 4.96MB 0% 0% 21 entries were displayed.
Storage Efficiency、ボリューム、aggregate、Snapshotの情報を確認します。
::*> volume show -volume vol1* -fields available, filesystem-size, total, used, percent-used, performance-tier-inactive-user-data, performance-tier-inactive-user-data-percent, size, dedupe-space-saved, dedupe-space-saved-percent, dedupe-space-shared,logical-used, logical-used-percent,logical-used-by-afs, logical-available vserver volume size available filesystem-size total used percent-used dedupe-space-saved dedupe-space-saved-percent dedupe-space-shared logical-used logical-used-percent logical-available logical-used-by-afs performance-tier-inactive-user-data performance-tier-inactive-user-data-percent ------- ------ ---- --------- --------------- ------- ------ ------------ ------------------ -------------------------- ------------------- ------------ -------------------- ----------------- ------------------- ----------------------------------- ------------------------------------------- svm vol1 16GB 6.85GB 16GB 15.20GB 8.35GB 54% 808.8MB 9% 808.8MB 9.14GB 60% - 9.14GB - - svm2 vol1_dst 7.54GB 1.10GB 7.54GB 7.16GB 6.06GB 84% 4.98GB 45% 3.00GB 11.01GB 154% - 8.05GB 0B 0% svm2 vol1_dst2 9.84GB 1.45GB 9.84GB 9.35GB 7.90GB 84% 4.53GB 36% 4.10GB 12.40GB 133% - 9.06GB 0B 0% svm3 vol1_dst2_dst 10.00GB 1.62GB 10.00GB 10.00GB 8.37GB 83% 4.68GB 36% 2.15GB 13.05GB 131% - 9.14GB 0B 0% svm3 vol1_dst_dst 4.90GB 896.4MB 4.90GB 4.90GB 4.03GB 82% 5.11GB 56% 2.21GB 9.14GB 186% - 9.07GB 0B 0% 5 entries were displayed. ::*> volume show-footprint -volume vol1_dst2_dst Vserver : svm3 Volume : vol1_dst2_dst Feature Used Used% -------------------------------- ---------- ----- Volume Data Footprint 8.37GB 1% Footprint in Performance Tier 8.46GB 100% Footprint in FSxFabricpoolObjectStore 0B 0% Volume Guarantee 0B 0% Flexible Volume Metadata 61.92MB 0% Delayed Frees 93.18MB 0% File Operation Metadata 4KB 0% Total Footprint 8.52GB 1% Footprint Data Reduction 3.73GB 0% Auto Adaptive Compression 3.73GB 0% Effective Total Footprint 4.80GB 1% ::*> aggr show-efficiency -instance Name of the Aggregate: aggr1 Node where Aggregate Resides: FsxId0ab6f9b00824a187c-01 Logical Size Used by Volumes, Clones, Snapshot Copies in the Aggregate: 205.5GB Total Physical Used: 22.52GB Total Storage Efficiency Ratio: 9.13:1 Total Data Reduction Logical Used Without Snapshots: 44.03GB Total Data Reduction Physical Used Without Snapshots: 14.81GB Total Data Reduction Efficiency Ratio Without Snapshots: 2.97:1 Total Data Reduction Logical Used without snapshots and flexclones: 44.03GB Total Data Reduction Physical Used without snapshots and flexclones: 14.81GB Total Data Reduction Efficiency Ratio without snapshots and flexclones: 2.97:1 Total Logical Size Used by Volumes, Clones, Snapshot Copies in the FabricPool Performance Tier: 150.9GB Total Physical Used in FabricPool Performance Tier: 18.86GB Total FabricPool Performance Tier Storage Efficiency Ratio: 8.00:1 Total Data Reduction Logical Used without snapshots and flexclones in the FabricPool Performance Tier: 35.65GB Total Data Reduction Physical Used without snapshots and flexclones in the FabricPool Performance Tier: 11.17GB Total FabricPool Performance Tier Data Reduction Efficiency Ratio without snapshots and flexclones: 3.19:1 Logical Space Used for All Volumes: 44.03GB Physical Space Used for All Volumes: 23.93GB Space Saved by Volume Deduplication: 20.09GB Space Saved by Volume Deduplication and pattern detection: 20.09GB Volume Deduplication Savings ratio: 1.84:1 Space Saved by Volume Compression: 0B Volume Compression Savings ratio: 1.00:1 Space Saved by Inline Zero Pattern Detection: 0B Volume Data Reduction SE Ratio: 1.84:1 Logical Space Used by the Aggregate: 32.67GB Physical Space Used by the Aggregate: 22.52GB Space Saved by Aggregate Data Reduction: 10.15GB Aggregate Data Reduction SE Ratio: 1.45:1 Logical Size Used by Snapshot Copies: 161.5GB Physical Size Used by Snapshot Copies: 11.18GB Snapshot Volume Data Reduction Ratio: 14.44:1 Logical Size Used by FlexClone Volumes: 0B Physical Sized Used by FlexClone Volumes: 0B FlexClone Volume Data Reduction Ratio: 1.00:1 Snapshot And FlexClone Volume Data Reduction SE Ratio: 14.44:1 Number of Volumes Offline: 0 Number of SIS Disabled Volumes: 3 Number of SIS Change Log Disabled Volumes: 0
やはり、特に変化はありませんね。
NFSクライアントから確認しても特に変化はありません。
$ df -hT -t nfs4 Filesystem Type Size Used Avail Use% Mounted on svm-0058ae83d258ab2e3.fs-0ab6f9b00824a187c.fsx.us-east-1.amazonaws.com:/vol1 nfs4 16G 8.4G 6.9G 55% /mnt/fsxn/vol1 svm-00c880c6c6eb922ed.fs-0ab6f9b00824a187c.fsx.us-east-1.amazonaws.com:/vol1_dst_dst nfs4 5.0G 4.1G 897M 83% /mnt/fsxn/vol1_dst_dst svm-00c880c6c6eb922ed.fs-0ab6f9b00824a187c.fsx.us-east-1.amazonaws.com:/vol1_dst2_dst nfs4 10G 8.4G 1.7G 84% /mnt/fsxn/vol1_dst2_dst
3.83GBのSnapshotを削除します。
::*> snapshot delete -vserver svm3 -volume vol1_dst2_dst -snapshot snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779291.2023-12-25_050406 Warning: Deleting a Snapshot copy permanently removes data that is stored only in that Snapshot copy. Are you sure you want to delete Snapshot copy "snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779291.2023-12-25_050406" for volume "vol1_dst2_dst" in Vserver "svm3" ? {y|n}: y ::*> snapshot show -volume vol1* ---Blocks--- Vserver Volume Snapshot Size Total% Used% -------- -------- ------------------------------------- -------- ------ ----- svm vol1 test.2023-12-22_0533 160KB 0% 0% snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_053528 24.45MB 0% 0% snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_065900 312KB 0% 0% snapmirror.674509fa-a065-11ee-981e-bdd56ead09c8_2148779290.2023-12-22_075507 292KB 0% 0% snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779291.2023-12-25_050406 148KB 0% 0% snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779291.2023-12-25_060027 136KB 0% 0% svm2 vol1_dst snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_015623 1.91GB 25% 38% test.2023-12-22_0533 893.2MB 12% 22% snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_053528 481.9MB 6% 13% snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_065900 82.71MB 1% 3% vol1_dst2 test.2023-12-22_0533 608KB 0% 0% snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_053528 364KB 0% 0% snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_065900 420KB 0% 0% snapmirror.674509fa-a065-11ee-981e-bdd56ead09c8_2148779290.2023-12-22_075507 280KB 0% 0% snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779291.2023-12-25_050406 3.83GB 39% 46% snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779291.2023-12-25_060027 140KB 0% 0% svm3 vol1_dst2_dst snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779291.2023-12-25_060027 94.08MB 2% 2% vol1_dst_dst snapmirror.5af907bb-a065-11ee-981e-bdd56ead09c8_2148779289.2023-12-22_065900 24.16MB 0% 1% snapmirror.674509fa-a065-11ee-981e-bdd56ead09c8_2148779290.2023-12-22_074801 41.57MB 1% 1% snapmirror.674509fa-a065-11ee-981e-bdd56ead09c8_2148779290.2023-12-22_075507 4.96MB 0% 0% 20 entries were displayed.
Storage Efficiency、ボリューム、aggregate、Snapshotの情報を確認します。
::*> volume show -volume vol1* -fields available, filesystem-size, total, used, percent-used, performance-tier-inactive-user-data, performance-tier-inactive-user-data-percent, size, dedupe-space-saved, dedupe-space-saved-percent, dedupe-space-shared,logical-used, logical-used-percent,logical-used-by-afs, logical-available vserver volume size available filesystem-size total used percent-used dedupe-space-saved dedupe-space-saved-percent dedupe-space-shared logical-used logical-used-percent logical-available logical-used-by-afs performance-tier-inactive-user-data performance-tier-inactive-user-data-percent ------- ------ ---- --------- --------------- ------- ------ ------------ ------------------ -------------------------- ------------------- ------------ -------------------- ----------------- ------------------- ----------------------------------- ------------------------------------------- svm vol1 16GB 6.85GB 16GB 15.20GB 8.35GB 54% 808.8MB 9% 808.8MB 9.14GB 60% - 9.14GB - - svm2 vol1_dst 7.54GB 1.10GB 7.54GB 7.16GB 6.06GB 84% 4.98GB 45% 3.00GB 11.01GB 154% - 8.05GB 0B 0% svm2 vol1_dst2 9.84GB 1.45GB 9.84GB 9.35GB 7.90GB 84% 4.53GB 36% 4.10GB 12.40GB 133% - 9.06GB 0B 0% svm3 vol1_dst2_dst 5.40GB 876.4MB 5.40GB 5.40GB 4.55GB 84% 4.68GB 51% 2.15GB 9.23GB 171% - 9.14GB 0B 0% svm3 vol1_dst_dst 4.90GB 896.4MB 4.90GB 4.90GB 4.03GB 82% 5.11GB 56% 2.21GB 9.14GB 186% - 9.07GB 0B 0% 5 entries were displayed. ::*> volume show-footprint -volume vol1_dst2_dst Vserver : svm3 Volume : vol1_dst2_dst Feature Used Used% -------------------------------- ---------- ----- Volume Data Footprint 4.55GB 1% Footprint in Performance Tier 7.42GB 100% Footprint in FSxFabricpoolObjectStore 0B 0% Volume Guarantee 0B 0% Flexible Volume Metadata 0B 0% Delayed Frees 2.87GB 0% File Operation Metadata 4KB 0% Total Footprint 7.42GB 1% Footprint Data Reduction 3.27GB 0% Auto Adaptive Compression 3.27GB 0% Effective Total Footprint 4.15GB 0% ::*> aggr show-efficiency -instance Name of the Aggregate: aggr1 Node where Aggregate Resides: FsxId0ab6f9b00824a187c-01 Logical Size Used by Volumes, Clones, Snapshot Copies in the Aggregate: 196.5GB Total Physical Used: 19.67GB Total Storage Efficiency Ratio: 9.99:1 Total Data Reduction Logical Used Without Snapshots: 44.03GB Total Data Reduction Physical Used Without Snapshots: 14.81GB Total Data Reduction Efficiency Ratio Without Snapshots: 2.97:1 Total Data Reduction Logical Used without snapshots and flexclones: 44.03GB Total Data Reduction Physical Used without snapshots and flexclones: 14.81GB Total Data Reduction Efficiency Ratio without snapshots and flexclones: 2.97:1 Total Logical Size Used by Volumes, Clones, Snapshot Copies in the FabricPool Performance Tier: 141.8GB Total Physical Used in FabricPool Performance Tier: 16.01GB Total FabricPool Performance Tier Storage Efficiency Ratio: 8.86:1 Total Data Reduction Logical Used without snapshots and flexclones in the FabricPool Performance Tier: 35.65GB Total Data Reduction Physical Used without snapshots and flexclones in the FabricPool Performance Tier: 11.17GB Total FabricPool Performance Tier Data Reduction Efficiency Ratio without snapshots and flexclones: 3.19:1 Logical Space Used for All Volumes: 44.03GB Physical Space Used for All Volumes: 23.93GB Space Saved by Volume Deduplication: 20.09GB Space Saved by Volume Deduplication and pattern detection: 20.09GB Volume Deduplication Savings ratio: 1.84:1 Space Saved by Volume Compression: 0B Volume Compression Savings ratio: 1.00:1 Space Saved by Inline Zero Pattern Detection: 0B Volume Data Reduction SE Ratio: 1.84:1 Logical Space Used by the Aggregate: 29.82GB Physical Space Used by the Aggregate: 19.67GB Space Saved by Aggregate Data Reduction: 10.15GB Aggregate Data Reduction SE Ratio: 1.52:1 Logical Size Used by Snapshot Copies: 152.5GB Physical Size Used by Snapshot Copies: 7.36GB Snapshot Volume Data Reduction Ratio: 20.72:1 Logical Size Used by FlexClone Volumes: 0B Physical Sized Used by FlexClone Volumes: 0B FlexClone Volume Data Reduction Ratio: 1.00:1 Snapshot And FlexClone Volume Data Reduction SE Ratio: 20.72:1 Number of Volumes Offline: 0 Number of SIS Disabled Volumes: 3 Number of SIS Change Log Disabled Volumes: 0
Snapshotを削除したことにより、以下のように各種値が変化していました。
used
: 8.37GB -> 4.55GBlogical-used
: 13.05GB -> 9.23GBTotal Physical Used
: 22.52GB -> 19.67GB
3.85GBのSnapshotを削除したのに応じてボリューム、aggregateの物理使用量が減っていることが分かりますね。
NFSクライアントからも確認します。
$ df -hT -t nfs4 Filesystem Type Size Used Avail Use% Mounted on svm-0058ae83d258ab2e3.fs-0ab6f9b00824a187c.fsx.us-east-1.amazonaws.com:/vol1 nfs4 16G 8.4G 6.9G 55% /mnt/fsxn/vol1 svm-00c880c6c6eb922ed.fs-0ab6f9b00824a187c.fsx.us-east-1.amazonaws.com:/vol1_dst_dst nfs4 5.0G 4.1G 897M 83% /mnt/fsxn/vol1_dst_dst svm-00c880c6c6eb922ed.fs-0ab6f9b00824a187c.fsx.us-east-1.amazonaws.com:/vol1_dst2_dst nfs4 5.5G 4.6G 878M 85% /mnt/fsxn/vol1_dst2_dst
使用量が8.4Gから4.6Gとなっていました。
移行や二次バックアップに
Amazon FSx for NetApp ONTAPでカスケードSnapMirrorを試してみました。
移行や二次バックアップに役立ちそうですね。特にオリジンの操作を行いたくない時には使っていきたいです。
この記事が誰かの助けになれば幸いです。
以上、AWS事業本部 コンサルティング部の のんピ(@non____97)でした!